ciao
This commit is contained in:
parent
a09f2f22b2
commit
f649c1acb7
File diff suppressed because it is too large
Load Diff
@ -1 +0,0 @@
|
|||||||
pip
|
|
@ -1,20 +0,0 @@
|
|||||||
Copyright (c) 2008-present The pip developers (see AUTHORS.txt file)
|
|
||||||
|
|
||||||
Permission is hereby granted, free of charge, to any person obtaining
|
|
||||||
a copy of this software and associated documentation files (the
|
|
||||||
"Software"), to deal in the Software without restriction, including
|
|
||||||
without limitation the rights to use, copy, modify, merge, publish,
|
|
||||||
distribute, sublicense, and/or sell copies of the Software, and to
|
|
||||||
permit persons to whom the Software is furnished to do so, subject to
|
|
||||||
the following conditions:
|
|
||||||
|
|
||||||
The above copyright notice and this permission notice shall be
|
|
||||||
included in all copies or substantial portions of the Software.
|
|
||||||
|
|
||||||
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
||||||
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
||||||
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
||||||
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
|
||||||
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
|
|
||||||
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
|
|
||||||
WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
|
@ -1,90 +0,0 @@
|
|||||||
Metadata-Version: 2.1
|
|
||||||
Name: pip
|
|
||||||
Version: 23.2.1
|
|
||||||
Summary: The PyPA recommended tool for installing Python packages.
|
|
||||||
Home-page: https://pip.pypa.io/
|
|
||||||
Author: The pip developers
|
|
||||||
Author-email: distutils-sig@python.org
|
|
||||||
License: MIT
|
|
||||||
Project-URL: Documentation, https://pip.pypa.io
|
|
||||||
Project-URL: Source, https://github.com/pypa/pip
|
|
||||||
Project-URL: Changelog, https://pip.pypa.io/en/stable/news/
|
|
||||||
Classifier: Development Status :: 5 - Production/Stable
|
|
||||||
Classifier: Intended Audience :: Developers
|
|
||||||
Classifier: License :: OSI Approved :: MIT License
|
|
||||||
Classifier: Topic :: Software Development :: Build Tools
|
|
||||||
Classifier: Programming Language :: Python
|
|
||||||
Classifier: Programming Language :: Python :: 3
|
|
||||||
Classifier: Programming Language :: Python :: 3 :: Only
|
|
||||||
Classifier: Programming Language :: Python :: 3.7
|
|
||||||
Classifier: Programming Language :: Python :: 3.8
|
|
||||||
Classifier: Programming Language :: Python :: 3.9
|
|
||||||
Classifier: Programming Language :: Python :: 3.10
|
|
||||||
Classifier: Programming Language :: Python :: 3.11
|
|
||||||
Classifier: Programming Language :: Python :: 3.12
|
|
||||||
Classifier: Programming Language :: Python :: Implementation :: CPython
|
|
||||||
Classifier: Programming Language :: Python :: Implementation :: PyPy
|
|
||||||
Requires-Python: >=3.7
|
|
||||||
License-File: LICENSE.txt
|
|
||||||
License-File: AUTHORS.txt
|
|
||||||
|
|
||||||
pip - The Python Package Installer
|
|
||||||
==================================
|
|
||||||
|
|
||||||
.. image:: https://img.shields.io/pypi/v/pip.svg
|
|
||||||
:target: https://pypi.org/project/pip/
|
|
||||||
|
|
||||||
.. image:: https://readthedocs.org/projects/pip/badge/?version=latest
|
|
||||||
:target: https://pip.pypa.io/en/latest
|
|
||||||
|
|
||||||
pip is the `package installer`_ for Python. You can use pip to install packages from the `Python Package Index`_ and other indexes.
|
|
||||||
|
|
||||||
Please take a look at our documentation for how to install and use pip:
|
|
||||||
|
|
||||||
* `Installation`_
|
|
||||||
* `Usage`_
|
|
||||||
|
|
||||||
We release updates regularly, with a new version every 3 months. Find more details in our documentation:
|
|
||||||
|
|
||||||
* `Release notes`_
|
|
||||||
* `Release process`_
|
|
||||||
|
|
||||||
In pip 20.3, we've `made a big improvement to the heart of pip`_; `learn more`_. We want your input, so `sign up for our user experience research studies`_ to help us do it right.
|
|
||||||
|
|
||||||
**Note**: pip 21.0, in January 2021, removed Python 2 support, per pip's `Python 2 support policy`_. Please migrate to Python 3.
|
|
||||||
|
|
||||||
If you find bugs, need help, or want to talk to the developers, please use our mailing lists or chat rooms:
|
|
||||||
|
|
||||||
* `Issue tracking`_
|
|
||||||
* `Discourse channel`_
|
|
||||||
* `User IRC`_
|
|
||||||
|
|
||||||
If you want to get involved head over to GitHub to get the source code, look at our development documentation and feel free to jump on the developer mailing lists and chat rooms:
|
|
||||||
|
|
||||||
* `GitHub page`_
|
|
||||||
* `Development documentation`_
|
|
||||||
* `Development IRC`_
|
|
||||||
|
|
||||||
Code of Conduct
|
|
||||||
---------------
|
|
||||||
|
|
||||||
Everyone interacting in the pip project's codebases, issue trackers, chat
|
|
||||||
rooms, and mailing lists is expected to follow the `PSF Code of Conduct`_.
|
|
||||||
|
|
||||||
.. _package installer: https://packaging.python.org/guides/tool-recommendations/
|
|
||||||
.. _Python Package Index: https://pypi.org
|
|
||||||
.. _Installation: https://pip.pypa.io/en/stable/installation/
|
|
||||||
.. _Usage: https://pip.pypa.io/en/stable/
|
|
||||||
.. _Release notes: https://pip.pypa.io/en/stable/news.html
|
|
||||||
.. _Release process: https://pip.pypa.io/en/latest/development/release-process/
|
|
||||||
.. _GitHub page: https://github.com/pypa/pip
|
|
||||||
.. _Development documentation: https://pip.pypa.io/en/latest/development
|
|
||||||
.. _made a big improvement to the heart of pip: https://pyfound.blogspot.com/2020/11/pip-20-3-new-resolver.html
|
|
||||||
.. _learn more: https://pip.pypa.io/en/latest/user_guide/#changes-to-the-pip-dependency-resolver-in-20-3-2020
|
|
||||||
.. _sign up for our user experience research studies: https://pyfound.blogspot.com/2020/03/new-pip-resolver-to-roll-out-this-year.html
|
|
||||||
.. _Python 2 support policy: https://pip.pypa.io/en/latest/development/release-process/#python-2-support
|
|
||||||
.. _Issue tracking: https://github.com/pypa/pip/issues
|
|
||||||
.. _Discourse channel: https://discuss.python.org/c/packaging
|
|
||||||
.. _User IRC: https://kiwiirc.com/nextclient/#ircs://irc.libera.chat:+6697/pypa
|
|
||||||
.. _Development IRC: https://kiwiirc.com/nextclient/#ircs://irc.libera.chat:+6697/pypa-dev
|
|
||||||
.. _PSF Code of Conduct: https://github.com/pypa/.github/blob/main/CODE_OF_CONDUCT.md
|
|
File diff suppressed because it is too large
Load Diff
@ -1,5 +0,0 @@
|
|||||||
Wheel-Version: 1.0
|
|
||||||
Generator: bdist_wheel (0.40.0)
|
|
||||||
Root-Is-Purelib: true
|
|
||||||
Tag: py3-none-any
|
|
||||||
|
|
@ -1,4 +0,0 @@
|
|||||||
[console_scripts]
|
|
||||||
pip = pip._internal.cli.main:main
|
|
||||||
pip3 = pip._internal.cli.main:main
|
|
||||||
pip3.11 = pip._internal.cli.main:main
|
|
@ -1 +0,0 @@
|
|||||||
pip
|
|
@ -1,13 +0,0 @@
|
|||||||
from typing import List, Optional
|
|
||||||
|
|
||||||
__version__ = "23.2.1"
|
|
||||||
|
|
||||||
|
|
||||||
def main(args: Optional[List[str]] = None) -> int:
|
|
||||||
"""This is an internal API only meant for use by pip's own console scripts.
|
|
||||||
|
|
||||||
For additional details, see https://github.com/pypa/pip/issues/7498.
|
|
||||||
"""
|
|
||||||
from pip._internal.utils.entrypoints import _wrapper
|
|
||||||
|
|
||||||
return _wrapper(args)
|
|
@ -1,24 +0,0 @@
|
|||||||
import os
|
|
||||||
import sys
|
|
||||||
|
|
||||||
# Remove '' and current working directory from the first entry
|
|
||||||
# of sys.path, if present to avoid using current directory
|
|
||||||
# in pip commands check, freeze, install, list and show,
|
|
||||||
# when invoked as python -m pip <command>
|
|
||||||
if sys.path[0] in ("", os.getcwd()):
|
|
||||||
sys.path.pop(0)
|
|
||||||
|
|
||||||
# If we are running from a wheel, add the wheel to sys.path
|
|
||||||
# This allows the usage python pip-*.whl/pip install pip-*.whl
|
|
||||||
if __package__ == "":
|
|
||||||
# __file__ is pip-*.whl/pip/__main__.py
|
|
||||||
# first dirname call strips of '/__main__.py', second strips off '/pip'
|
|
||||||
# Resulting path is the name of the wheel itself
|
|
||||||
# Add that to sys.path so we can import pip
|
|
||||||
path = os.path.dirname(os.path.dirname(__file__))
|
|
||||||
sys.path.insert(0, path)
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
from pip._internal.cli.main import main as _main
|
|
||||||
|
|
||||||
sys.exit(_main())
|
|
@ -1,50 +0,0 @@
|
|||||||
"""Execute exactly this copy of pip, within a different environment.
|
|
||||||
|
|
||||||
This file is named as it is, to ensure that this module can't be imported via
|
|
||||||
an import statement.
|
|
||||||
"""
|
|
||||||
|
|
||||||
# /!\ This version compatibility check section must be Python 2 compatible. /!\
|
|
||||||
|
|
||||||
import sys
|
|
||||||
|
|
||||||
# Copied from setup.py
|
|
||||||
PYTHON_REQUIRES = (3, 7)
|
|
||||||
|
|
||||||
|
|
||||||
def version_str(version): # type: ignore
|
|
||||||
return ".".join(str(v) for v in version)
|
|
||||||
|
|
||||||
|
|
||||||
if sys.version_info[:2] < PYTHON_REQUIRES:
|
|
||||||
raise SystemExit(
|
|
||||||
"This version of pip does not support python {} (requires >={}).".format(
|
|
||||||
version_str(sys.version_info[:2]), version_str(PYTHON_REQUIRES)
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
# From here on, we can use Python 3 features, but the syntax must remain
|
|
||||||
# Python 2 compatible.
|
|
||||||
|
|
||||||
import runpy # noqa: E402
|
|
||||||
from importlib.machinery import PathFinder # noqa: E402
|
|
||||||
from os.path import dirname # noqa: E402
|
|
||||||
|
|
||||||
PIP_SOURCES_ROOT = dirname(dirname(__file__))
|
|
||||||
|
|
||||||
|
|
||||||
class PipImportRedirectingFinder:
|
|
||||||
@classmethod
|
|
||||||
def find_spec(self, fullname, path=None, target=None): # type: ignore
|
|
||||||
if fullname != "pip":
|
|
||||||
return None
|
|
||||||
|
|
||||||
spec = PathFinder.find_spec(fullname, [PIP_SOURCES_ROOT], target)
|
|
||||||
assert spec, (PIP_SOURCES_ROOT, fullname)
|
|
||||||
return spec
|
|
||||||
|
|
||||||
|
|
||||||
sys.meta_path.insert(0, PipImportRedirectingFinder())
|
|
||||||
|
|
||||||
assert __name__ == "__main__", "Cannot run __pip-runner__.py as a non-main module"
|
|
||||||
runpy.run_module("pip", run_name="__main__", alter_sys=True)
|
|
Binary file not shown.
Binary file not shown.
Binary file not shown.
@ -1,19 +0,0 @@
|
|||||||
from typing import List, Optional
|
|
||||||
|
|
||||||
import pip._internal.utils.inject_securetransport # noqa
|
|
||||||
from pip._internal.utils import _log
|
|
||||||
|
|
||||||
# init_logging() must be called before any call to logging.getLogger()
|
|
||||||
# which happens at import of most modules.
|
|
||||||
_log.init_logging()
|
|
||||||
|
|
||||||
|
|
||||||
def main(args: (Optional[List[str]]) = None) -> int:
|
|
||||||
"""This is preserved for old console scripts that may still be referencing
|
|
||||||
it.
|
|
||||||
|
|
||||||
For additional details, see https://github.com/pypa/pip/issues/7498.
|
|
||||||
"""
|
|
||||||
from pip._internal.utils.entrypoints import _wrapper
|
|
||||||
|
|
||||||
return _wrapper(args)
|
|
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
@ -1,311 +0,0 @@
|
|||||||
"""Build Environment used for isolation during sdist building
|
|
||||||
"""
|
|
||||||
|
|
||||||
import logging
|
|
||||||
import os
|
|
||||||
import pathlib
|
|
||||||
import site
|
|
||||||
import sys
|
|
||||||
import textwrap
|
|
||||||
from collections import OrderedDict
|
|
||||||
from types import TracebackType
|
|
||||||
from typing import TYPE_CHECKING, Iterable, List, Optional, Set, Tuple, Type, Union
|
|
||||||
|
|
||||||
from pip._vendor.certifi import where
|
|
||||||
from pip._vendor.packaging.requirements import Requirement
|
|
||||||
from pip._vendor.packaging.version import Version
|
|
||||||
|
|
||||||
from pip import __file__ as pip_location
|
|
||||||
from pip._internal.cli.spinners import open_spinner
|
|
||||||
from pip._internal.locations import get_platlib, get_purelib, get_scheme
|
|
||||||
from pip._internal.metadata import get_default_environment, get_environment
|
|
||||||
from pip._internal.utils.subprocess import call_subprocess
|
|
||||||
from pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from pip._internal.index.package_finder import PackageFinder
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def _dedup(a: str, b: str) -> Union[Tuple[str], Tuple[str, str]]:
|
|
||||||
return (a, b) if a != b else (a,)
|
|
||||||
|
|
||||||
|
|
||||||
class _Prefix:
|
|
||||||
def __init__(self, path: str) -> None:
|
|
||||||
self.path = path
|
|
||||||
self.setup = False
|
|
||||||
scheme = get_scheme("", prefix=path)
|
|
||||||
self.bin_dir = scheme.scripts
|
|
||||||
self.lib_dirs = _dedup(scheme.purelib, scheme.platlib)
|
|
||||||
|
|
||||||
|
|
||||||
def get_runnable_pip() -> str:
|
|
||||||
"""Get a file to pass to a Python executable, to run the currently-running pip.
|
|
||||||
|
|
||||||
This is used to run a pip subprocess, for installing requirements into the build
|
|
||||||
environment.
|
|
||||||
"""
|
|
||||||
source = pathlib.Path(pip_location).resolve().parent
|
|
||||||
|
|
||||||
if not source.is_dir():
|
|
||||||
# This would happen if someone is using pip from inside a zip file. In that
|
|
||||||
# case, we can use that directly.
|
|
||||||
return str(source)
|
|
||||||
|
|
||||||
return os.fsdecode(source / "__pip-runner__.py")
|
|
||||||
|
|
||||||
|
|
||||||
def _get_system_sitepackages() -> Set[str]:
|
|
||||||
"""Get system site packages
|
|
||||||
|
|
||||||
Usually from site.getsitepackages,
|
|
||||||
but fallback on `get_purelib()/get_platlib()` if unavailable
|
|
||||||
(e.g. in a virtualenv created by virtualenv<20)
|
|
||||||
|
|
||||||
Returns normalized set of strings.
|
|
||||||
"""
|
|
||||||
if hasattr(site, "getsitepackages"):
|
|
||||||
system_sites = site.getsitepackages()
|
|
||||||
else:
|
|
||||||
# virtualenv < 20 overwrites site.py without getsitepackages
|
|
||||||
# fallback on get_purelib/get_platlib.
|
|
||||||
# this is known to miss things, but shouldn't in the cases
|
|
||||||
# where getsitepackages() has been removed (inside a virtualenv)
|
|
||||||
system_sites = [get_purelib(), get_platlib()]
|
|
||||||
return {os.path.normcase(path) for path in system_sites}
|
|
||||||
|
|
||||||
|
|
||||||
class BuildEnvironment:
|
|
||||||
"""Creates and manages an isolated environment to install build deps"""
|
|
||||||
|
|
||||||
def __init__(self) -> None:
|
|
||||||
temp_dir = TempDirectory(kind=tempdir_kinds.BUILD_ENV, globally_managed=True)
|
|
||||||
|
|
||||||
self._prefixes = OrderedDict(
|
|
||||||
(name, _Prefix(os.path.join(temp_dir.path, name)))
|
|
||||||
for name in ("normal", "overlay")
|
|
||||||
)
|
|
||||||
|
|
||||||
self._bin_dirs: List[str] = []
|
|
||||||
self._lib_dirs: List[str] = []
|
|
||||||
for prefix in reversed(list(self._prefixes.values())):
|
|
||||||
self._bin_dirs.append(prefix.bin_dir)
|
|
||||||
self._lib_dirs.extend(prefix.lib_dirs)
|
|
||||||
|
|
||||||
# Customize site to:
|
|
||||||
# - ensure .pth files are honored
|
|
||||||
# - prevent access to system site packages
|
|
||||||
system_sites = _get_system_sitepackages()
|
|
||||||
|
|
||||||
self._site_dir = os.path.join(temp_dir.path, "site")
|
|
||||||
if not os.path.exists(self._site_dir):
|
|
||||||
os.mkdir(self._site_dir)
|
|
||||||
with open(
|
|
||||||
os.path.join(self._site_dir, "sitecustomize.py"), "w", encoding="utf-8"
|
|
||||||
) as fp:
|
|
||||||
fp.write(
|
|
||||||
textwrap.dedent(
|
|
||||||
"""
|
|
||||||
import os, site, sys
|
|
||||||
|
|
||||||
# First, drop system-sites related paths.
|
|
||||||
original_sys_path = sys.path[:]
|
|
||||||
known_paths = set()
|
|
||||||
for path in {system_sites!r}:
|
|
||||||
site.addsitedir(path, known_paths=known_paths)
|
|
||||||
system_paths = set(
|
|
||||||
os.path.normcase(path)
|
|
||||||
for path in sys.path[len(original_sys_path):]
|
|
||||||
)
|
|
||||||
original_sys_path = [
|
|
||||||
path for path in original_sys_path
|
|
||||||
if os.path.normcase(path) not in system_paths
|
|
||||||
]
|
|
||||||
sys.path = original_sys_path
|
|
||||||
|
|
||||||
# Second, add lib directories.
|
|
||||||
# ensuring .pth file are processed.
|
|
||||||
for path in {lib_dirs!r}:
|
|
||||||
assert not path in sys.path
|
|
||||||
site.addsitedir(path)
|
|
||||||
"""
|
|
||||||
).format(system_sites=system_sites, lib_dirs=self._lib_dirs)
|
|
||||||
)
|
|
||||||
|
|
||||||
def __enter__(self) -> None:
|
|
||||||
self._save_env = {
|
|
||||||
name: os.environ.get(name, None)
|
|
||||||
for name in ("PATH", "PYTHONNOUSERSITE", "PYTHONPATH")
|
|
||||||
}
|
|
||||||
|
|
||||||
path = self._bin_dirs[:]
|
|
||||||
old_path = self._save_env["PATH"]
|
|
||||||
if old_path:
|
|
||||||
path.extend(old_path.split(os.pathsep))
|
|
||||||
|
|
||||||
pythonpath = [self._site_dir]
|
|
||||||
|
|
||||||
os.environ.update(
|
|
||||||
{
|
|
||||||
"PATH": os.pathsep.join(path),
|
|
||||||
"PYTHONNOUSERSITE": "1",
|
|
||||||
"PYTHONPATH": os.pathsep.join(pythonpath),
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
def __exit__(
|
|
||||||
self,
|
|
||||||
exc_type: Optional[Type[BaseException]],
|
|
||||||
exc_val: Optional[BaseException],
|
|
||||||
exc_tb: Optional[TracebackType],
|
|
||||||
) -> None:
|
|
||||||
for varname, old_value in self._save_env.items():
|
|
||||||
if old_value is None:
|
|
||||||
os.environ.pop(varname, None)
|
|
||||||
else:
|
|
||||||
os.environ[varname] = old_value
|
|
||||||
|
|
||||||
def check_requirements(
|
|
||||||
self, reqs: Iterable[str]
|
|
||||||
) -> Tuple[Set[Tuple[str, str]], Set[str]]:
|
|
||||||
"""Return 2 sets:
|
|
||||||
- conflicting requirements: set of (installed, wanted) reqs tuples
|
|
||||||
- missing requirements: set of reqs
|
|
||||||
"""
|
|
||||||
missing = set()
|
|
||||||
conflicting = set()
|
|
||||||
if reqs:
|
|
||||||
env = (
|
|
||||||
get_environment(self._lib_dirs)
|
|
||||||
if hasattr(self, "_lib_dirs")
|
|
||||||
else get_default_environment()
|
|
||||||
)
|
|
||||||
for req_str in reqs:
|
|
||||||
req = Requirement(req_str)
|
|
||||||
# We're explicitly evaluating with an empty extra value, since build
|
|
||||||
# environments are not provided any mechanism to select specific extras.
|
|
||||||
if req.marker is not None and not req.marker.evaluate({"extra": ""}):
|
|
||||||
continue
|
|
||||||
dist = env.get_distribution(req.name)
|
|
||||||
if not dist:
|
|
||||||
missing.add(req_str)
|
|
||||||
continue
|
|
||||||
if isinstance(dist.version, Version):
|
|
||||||
installed_req_str = f"{req.name}=={dist.version}"
|
|
||||||
else:
|
|
||||||
installed_req_str = f"{req.name}==={dist.version}"
|
|
||||||
if not req.specifier.contains(dist.version, prereleases=True):
|
|
||||||
conflicting.add((installed_req_str, req_str))
|
|
||||||
# FIXME: Consider direct URL?
|
|
||||||
return conflicting, missing
|
|
||||||
|
|
||||||
def install_requirements(
|
|
||||||
self,
|
|
||||||
finder: "PackageFinder",
|
|
||||||
requirements: Iterable[str],
|
|
||||||
prefix_as_string: str,
|
|
||||||
*,
|
|
||||||
kind: str,
|
|
||||||
) -> None:
|
|
||||||
prefix = self._prefixes[prefix_as_string]
|
|
||||||
assert not prefix.setup
|
|
||||||
prefix.setup = True
|
|
||||||
if not requirements:
|
|
||||||
return
|
|
||||||
self._install_requirements(
|
|
||||||
get_runnable_pip(),
|
|
||||||
finder,
|
|
||||||
requirements,
|
|
||||||
prefix,
|
|
||||||
kind=kind,
|
|
||||||
)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _install_requirements(
|
|
||||||
pip_runnable: str,
|
|
||||||
finder: "PackageFinder",
|
|
||||||
requirements: Iterable[str],
|
|
||||||
prefix: _Prefix,
|
|
||||||
*,
|
|
||||||
kind: str,
|
|
||||||
) -> None:
|
|
||||||
args: List[str] = [
|
|
||||||
sys.executable,
|
|
||||||
pip_runnable,
|
|
||||||
"install",
|
|
||||||
"--ignore-installed",
|
|
||||||
"--no-user",
|
|
||||||
"--prefix",
|
|
||||||
prefix.path,
|
|
||||||
"--no-warn-script-location",
|
|
||||||
]
|
|
||||||
if logger.getEffectiveLevel() <= logging.DEBUG:
|
|
||||||
args.append("-v")
|
|
||||||
for format_control in ("no_binary", "only_binary"):
|
|
||||||
formats = getattr(finder.format_control, format_control)
|
|
||||||
args.extend(
|
|
||||||
(
|
|
||||||
"--" + format_control.replace("_", "-"),
|
|
||||||
",".join(sorted(formats or {":none:"})),
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
index_urls = finder.index_urls
|
|
||||||
if index_urls:
|
|
||||||
args.extend(["-i", index_urls[0]])
|
|
||||||
for extra_index in index_urls[1:]:
|
|
||||||
args.extend(["--extra-index-url", extra_index])
|
|
||||||
else:
|
|
||||||
args.append("--no-index")
|
|
||||||
for link in finder.find_links:
|
|
||||||
args.extend(["--find-links", link])
|
|
||||||
|
|
||||||
for host in finder.trusted_hosts:
|
|
||||||
args.extend(["--trusted-host", host])
|
|
||||||
if finder.allow_all_prereleases:
|
|
||||||
args.append("--pre")
|
|
||||||
if finder.prefer_binary:
|
|
||||||
args.append("--prefer-binary")
|
|
||||||
args.append("--")
|
|
||||||
args.extend(requirements)
|
|
||||||
extra_environ = {"_PIP_STANDALONE_CERT": where()}
|
|
||||||
with open_spinner(f"Installing {kind}") as spinner:
|
|
||||||
call_subprocess(
|
|
||||||
args,
|
|
||||||
command_desc=f"pip subprocess to install {kind}",
|
|
||||||
spinner=spinner,
|
|
||||||
extra_environ=extra_environ,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class NoOpBuildEnvironment(BuildEnvironment):
|
|
||||||
"""A no-op drop-in replacement for BuildEnvironment"""
|
|
||||||
|
|
||||||
def __init__(self) -> None:
|
|
||||||
pass
|
|
||||||
|
|
||||||
def __enter__(self) -> None:
|
|
||||||
pass
|
|
||||||
|
|
||||||
def __exit__(
|
|
||||||
self,
|
|
||||||
exc_type: Optional[Type[BaseException]],
|
|
||||||
exc_val: Optional[BaseException],
|
|
||||||
exc_tb: Optional[TracebackType],
|
|
||||||
) -> None:
|
|
||||||
pass
|
|
||||||
|
|
||||||
def cleanup(self) -> None:
|
|
||||||
pass
|
|
||||||
|
|
||||||
def install_requirements(
|
|
||||||
self,
|
|
||||||
finder: "PackageFinder",
|
|
||||||
requirements: Iterable[str],
|
|
||||||
prefix_as_string: str,
|
|
||||||
*,
|
|
||||||
kind: str,
|
|
||||||
) -> None:
|
|
||||||
raise NotImplementedError()
|
|
@ -1,292 +0,0 @@
|
|||||||
"""Cache Management
|
|
||||||
"""
|
|
||||||
|
|
||||||
import hashlib
|
|
||||||
import json
|
|
||||||
import logging
|
|
||||||
import os
|
|
||||||
from pathlib import Path
|
|
||||||
from typing import Any, Dict, List, Optional
|
|
||||||
|
|
||||||
from pip._vendor.packaging.tags import Tag, interpreter_name, interpreter_version
|
|
||||||
from pip._vendor.packaging.utils import canonicalize_name
|
|
||||||
|
|
||||||
from pip._internal.exceptions import InvalidWheelFilename
|
|
||||||
from pip._internal.models.direct_url import DirectUrl
|
|
||||||
from pip._internal.models.link import Link
|
|
||||||
from pip._internal.models.wheel import Wheel
|
|
||||||
from pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds
|
|
||||||
from pip._internal.utils.urls import path_to_url
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
ORIGIN_JSON_NAME = "origin.json"
|
|
||||||
|
|
||||||
|
|
||||||
def _hash_dict(d: Dict[str, str]) -> str:
|
|
||||||
"""Return a stable sha224 of a dictionary."""
|
|
||||||
s = json.dumps(d, sort_keys=True, separators=(",", ":"), ensure_ascii=True)
|
|
||||||
return hashlib.sha224(s.encode("ascii")).hexdigest()
|
|
||||||
|
|
||||||
|
|
||||||
class Cache:
|
|
||||||
"""An abstract class - provides cache directories for data from links
|
|
||||||
|
|
||||||
:param cache_dir: The root of the cache.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, cache_dir: str) -> None:
|
|
||||||
super().__init__()
|
|
||||||
assert not cache_dir or os.path.isabs(cache_dir)
|
|
||||||
self.cache_dir = cache_dir or None
|
|
||||||
|
|
||||||
def _get_cache_path_parts(self, link: Link) -> List[str]:
|
|
||||||
"""Get parts of part that must be os.path.joined with cache_dir"""
|
|
||||||
|
|
||||||
# We want to generate an url to use as our cache key, we don't want to
|
|
||||||
# just re-use the URL because it might have other items in the fragment
|
|
||||||
# and we don't care about those.
|
|
||||||
key_parts = {"url": link.url_without_fragment}
|
|
||||||
if link.hash_name is not None and link.hash is not None:
|
|
||||||
key_parts[link.hash_name] = link.hash
|
|
||||||
if link.subdirectory_fragment:
|
|
||||||
key_parts["subdirectory"] = link.subdirectory_fragment
|
|
||||||
|
|
||||||
# Include interpreter name, major and minor version in cache key
|
|
||||||
# to cope with ill-behaved sdists that build a different wheel
|
|
||||||
# depending on the python version their setup.py is being run on,
|
|
||||||
# and don't encode the difference in compatibility tags.
|
|
||||||
# https://github.com/pypa/pip/issues/7296
|
|
||||||
key_parts["interpreter_name"] = interpreter_name()
|
|
||||||
key_parts["interpreter_version"] = interpreter_version()
|
|
||||||
|
|
||||||
# Encode our key url with sha224, we'll use this because it has similar
|
|
||||||
# security properties to sha256, but with a shorter total output (and
|
|
||||||
# thus less secure). However the differences don't make a lot of
|
|
||||||
# difference for our use case here.
|
|
||||||
hashed = _hash_dict(key_parts)
|
|
||||||
|
|
||||||
# We want to nest the directories some to prevent having a ton of top
|
|
||||||
# level directories where we might run out of sub directories on some
|
|
||||||
# FS.
|
|
||||||
parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]]
|
|
||||||
|
|
||||||
return parts
|
|
||||||
|
|
||||||
def _get_candidates(self, link: Link, canonical_package_name: str) -> List[Any]:
|
|
||||||
can_not_cache = not self.cache_dir or not canonical_package_name or not link
|
|
||||||
if can_not_cache:
|
|
||||||
return []
|
|
||||||
|
|
||||||
candidates = []
|
|
||||||
path = self.get_path_for_link(link)
|
|
||||||
if os.path.isdir(path):
|
|
||||||
for candidate in os.listdir(path):
|
|
||||||
candidates.append((candidate, path))
|
|
||||||
return candidates
|
|
||||||
|
|
||||||
def get_path_for_link(self, link: Link) -> str:
|
|
||||||
"""Return a directory to store cached items in for link."""
|
|
||||||
raise NotImplementedError()
|
|
||||||
|
|
||||||
def get(
|
|
||||||
self,
|
|
||||||
link: Link,
|
|
||||||
package_name: Optional[str],
|
|
||||||
supported_tags: List[Tag],
|
|
||||||
) -> Link:
|
|
||||||
"""Returns a link to a cached item if it exists, otherwise returns the
|
|
||||||
passed link.
|
|
||||||
"""
|
|
||||||
raise NotImplementedError()
|
|
||||||
|
|
||||||
|
|
||||||
class SimpleWheelCache(Cache):
|
|
||||||
"""A cache of wheels for future installs."""
|
|
||||||
|
|
||||||
def __init__(self, cache_dir: str) -> None:
|
|
||||||
super().__init__(cache_dir)
|
|
||||||
|
|
||||||
def get_path_for_link(self, link: Link) -> str:
|
|
||||||
"""Return a directory to store cached wheels for link
|
|
||||||
|
|
||||||
Because there are M wheels for any one sdist, we provide a directory
|
|
||||||
to cache them in, and then consult that directory when looking up
|
|
||||||
cache hits.
|
|
||||||
|
|
||||||
We only insert things into the cache if they have plausible version
|
|
||||||
numbers, so that we don't contaminate the cache with things that were
|
|
||||||
not unique. E.g. ./package might have dozens of installs done for it
|
|
||||||
and build a version of 0.0...and if we built and cached a wheel, we'd
|
|
||||||
end up using the same wheel even if the source has been edited.
|
|
||||||
|
|
||||||
:param link: The link of the sdist for which this will cache wheels.
|
|
||||||
"""
|
|
||||||
parts = self._get_cache_path_parts(link)
|
|
||||||
assert self.cache_dir
|
|
||||||
# Store wheels within the root cache_dir
|
|
||||||
return os.path.join(self.cache_dir, "wheels", *parts)
|
|
||||||
|
|
||||||
def get(
|
|
||||||
self,
|
|
||||||
link: Link,
|
|
||||||
package_name: Optional[str],
|
|
||||||
supported_tags: List[Tag],
|
|
||||||
) -> Link:
|
|
||||||
candidates = []
|
|
||||||
|
|
||||||
if not package_name:
|
|
||||||
return link
|
|
||||||
|
|
||||||
canonical_package_name = canonicalize_name(package_name)
|
|
||||||
for wheel_name, wheel_dir in self._get_candidates(link, canonical_package_name):
|
|
||||||
try:
|
|
||||||
wheel = Wheel(wheel_name)
|
|
||||||
except InvalidWheelFilename:
|
|
||||||
continue
|
|
||||||
if canonicalize_name(wheel.name) != canonical_package_name:
|
|
||||||
logger.debug(
|
|
||||||
"Ignoring cached wheel %s for %s as it "
|
|
||||||
"does not match the expected distribution name %s.",
|
|
||||||
wheel_name,
|
|
||||||
link,
|
|
||||||
package_name,
|
|
||||||
)
|
|
||||||
continue
|
|
||||||
if not wheel.supported(supported_tags):
|
|
||||||
# Built for a different python/arch/etc
|
|
||||||
continue
|
|
||||||
candidates.append(
|
|
||||||
(
|
|
||||||
wheel.support_index_min(supported_tags),
|
|
||||||
wheel_name,
|
|
||||||
wheel_dir,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
if not candidates:
|
|
||||||
return link
|
|
||||||
|
|
||||||
_, wheel_name, wheel_dir = min(candidates)
|
|
||||||
return Link(path_to_url(os.path.join(wheel_dir, wheel_name)))
|
|
||||||
|
|
||||||
|
|
||||||
class EphemWheelCache(SimpleWheelCache):
|
|
||||||
"""A SimpleWheelCache that creates it's own temporary cache directory"""
|
|
||||||
|
|
||||||
def __init__(self) -> None:
|
|
||||||
self._temp_dir = TempDirectory(
|
|
||||||
kind=tempdir_kinds.EPHEM_WHEEL_CACHE,
|
|
||||||
globally_managed=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
super().__init__(self._temp_dir.path)
|
|
||||||
|
|
||||||
|
|
||||||
class CacheEntry:
|
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
link: Link,
|
|
||||||
persistent: bool,
|
|
||||||
):
|
|
||||||
self.link = link
|
|
||||||
self.persistent = persistent
|
|
||||||
self.origin: Optional[DirectUrl] = None
|
|
||||||
origin_direct_url_path = Path(self.link.file_path).parent / ORIGIN_JSON_NAME
|
|
||||||
if origin_direct_url_path.exists():
|
|
||||||
try:
|
|
||||||
self.origin = DirectUrl.from_json(
|
|
||||||
origin_direct_url_path.read_text(encoding="utf-8")
|
|
||||||
)
|
|
||||||
except Exception as e:
|
|
||||||
logger.warning(
|
|
||||||
"Ignoring invalid cache entry origin file %s for %s (%s)",
|
|
||||||
origin_direct_url_path,
|
|
||||||
link.filename,
|
|
||||||
e,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class WheelCache(Cache):
|
|
||||||
"""Wraps EphemWheelCache and SimpleWheelCache into a single Cache
|
|
||||||
|
|
||||||
This Cache allows for gracefully degradation, using the ephem wheel cache
|
|
||||||
when a certain link is not found in the simple wheel cache first.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, cache_dir: str) -> None:
|
|
||||||
super().__init__(cache_dir)
|
|
||||||
self._wheel_cache = SimpleWheelCache(cache_dir)
|
|
||||||
self._ephem_cache = EphemWheelCache()
|
|
||||||
|
|
||||||
def get_path_for_link(self, link: Link) -> str:
|
|
||||||
return self._wheel_cache.get_path_for_link(link)
|
|
||||||
|
|
||||||
def get_ephem_path_for_link(self, link: Link) -> str:
|
|
||||||
return self._ephem_cache.get_path_for_link(link)
|
|
||||||
|
|
||||||
def get(
|
|
||||||
self,
|
|
||||||
link: Link,
|
|
||||||
package_name: Optional[str],
|
|
||||||
supported_tags: List[Tag],
|
|
||||||
) -> Link:
|
|
||||||
cache_entry = self.get_cache_entry(link, package_name, supported_tags)
|
|
||||||
if cache_entry is None:
|
|
||||||
return link
|
|
||||||
return cache_entry.link
|
|
||||||
|
|
||||||
def get_cache_entry(
|
|
||||||
self,
|
|
||||||
link: Link,
|
|
||||||
package_name: Optional[str],
|
|
||||||
supported_tags: List[Tag],
|
|
||||||
) -> Optional[CacheEntry]:
|
|
||||||
"""Returns a CacheEntry with a link to a cached item if it exists or
|
|
||||||
None. The cache entry indicates if the item was found in the persistent
|
|
||||||
or ephemeral cache.
|
|
||||||
"""
|
|
||||||
retval = self._wheel_cache.get(
|
|
||||||
link=link,
|
|
||||||
package_name=package_name,
|
|
||||||
supported_tags=supported_tags,
|
|
||||||
)
|
|
||||||
if retval is not link:
|
|
||||||
return CacheEntry(retval, persistent=True)
|
|
||||||
|
|
||||||
retval = self._ephem_cache.get(
|
|
||||||
link=link,
|
|
||||||
package_name=package_name,
|
|
||||||
supported_tags=supported_tags,
|
|
||||||
)
|
|
||||||
if retval is not link:
|
|
||||||
return CacheEntry(retval, persistent=False)
|
|
||||||
|
|
||||||
return None
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def record_download_origin(cache_dir: str, download_info: DirectUrl) -> None:
|
|
||||||
origin_path = Path(cache_dir) / ORIGIN_JSON_NAME
|
|
||||||
if origin_path.exists():
|
|
||||||
try:
|
|
||||||
origin = DirectUrl.from_json(origin_path.read_text(encoding="utf-8"))
|
|
||||||
except Exception as e:
|
|
||||||
logger.warning(
|
|
||||||
"Could not read origin file %s in cache entry (%s). "
|
|
||||||
"Will attempt to overwrite it.",
|
|
||||||
origin_path,
|
|
||||||
e,
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
# TODO: use DirectUrl.equivalent when
|
|
||||||
# https://github.com/pypa/pip/pull/10564 is merged.
|
|
||||||
if origin.url != download_info.url:
|
|
||||||
logger.warning(
|
|
||||||
"Origin URL %s in cache entry %s does not match download URL "
|
|
||||||
"%s. This is likely a pip bug or a cache corruption issue. "
|
|
||||||
"Will overwrite it with the new value.",
|
|
||||||
origin.url,
|
|
||||||
cache_dir,
|
|
||||||
download_info.url,
|
|
||||||
)
|
|
||||||
origin_path.write_text(download_info.to_json(), encoding="utf-8")
|
|
@ -1,4 +0,0 @@
|
|||||||
"""Subpackage containing all of pip's command line interface related code
|
|
||||||
"""
|
|
||||||
|
|
||||||
# This file intentionally does not import submodules
|
|
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
@ -1,171 +0,0 @@
|
|||||||
"""Logic that powers autocompletion installed by ``pip completion``.
|
|
||||||
"""
|
|
||||||
|
|
||||||
import optparse
|
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
from itertools import chain
|
|
||||||
from typing import Any, Iterable, List, Optional
|
|
||||||
|
|
||||||
from pip._internal.cli.main_parser import create_main_parser
|
|
||||||
from pip._internal.commands import commands_dict, create_command
|
|
||||||
from pip._internal.metadata import get_default_environment
|
|
||||||
|
|
||||||
|
|
||||||
def autocomplete() -> None:
|
|
||||||
"""Entry Point for completion of main and subcommand options."""
|
|
||||||
# Don't complete if user hasn't sourced bash_completion file.
|
|
||||||
if "PIP_AUTO_COMPLETE" not in os.environ:
|
|
||||||
return
|
|
||||||
cwords = os.environ["COMP_WORDS"].split()[1:]
|
|
||||||
cword = int(os.environ["COMP_CWORD"])
|
|
||||||
try:
|
|
||||||
current = cwords[cword - 1]
|
|
||||||
except IndexError:
|
|
||||||
current = ""
|
|
||||||
|
|
||||||
parser = create_main_parser()
|
|
||||||
subcommands = list(commands_dict)
|
|
||||||
options = []
|
|
||||||
|
|
||||||
# subcommand
|
|
||||||
subcommand_name: Optional[str] = None
|
|
||||||
for word in cwords:
|
|
||||||
if word in subcommands:
|
|
||||||
subcommand_name = word
|
|
||||||
break
|
|
||||||
# subcommand options
|
|
||||||
if subcommand_name is not None:
|
|
||||||
# special case: 'help' subcommand has no options
|
|
||||||
if subcommand_name == "help":
|
|
||||||
sys.exit(1)
|
|
||||||
# special case: list locally installed dists for show and uninstall
|
|
||||||
should_list_installed = not current.startswith("-") and subcommand_name in [
|
|
||||||
"show",
|
|
||||||
"uninstall",
|
|
||||||
]
|
|
||||||
if should_list_installed:
|
|
||||||
env = get_default_environment()
|
|
||||||
lc = current.lower()
|
|
||||||
installed = [
|
|
||||||
dist.canonical_name
|
|
||||||
for dist in env.iter_installed_distributions(local_only=True)
|
|
||||||
if dist.canonical_name.startswith(lc)
|
|
||||||
and dist.canonical_name not in cwords[1:]
|
|
||||||
]
|
|
||||||
# if there are no dists installed, fall back to option completion
|
|
||||||
if installed:
|
|
||||||
for dist in installed:
|
|
||||||
print(dist)
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
should_list_installables = (
|
|
||||||
not current.startswith("-") and subcommand_name == "install"
|
|
||||||
)
|
|
||||||
if should_list_installables:
|
|
||||||
for path in auto_complete_paths(current, "path"):
|
|
||||||
print(path)
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
subcommand = create_command(subcommand_name)
|
|
||||||
|
|
||||||
for opt in subcommand.parser.option_list_all:
|
|
||||||
if opt.help != optparse.SUPPRESS_HELP:
|
|
||||||
for opt_str in opt._long_opts + opt._short_opts:
|
|
||||||
options.append((opt_str, opt.nargs))
|
|
||||||
|
|
||||||
# filter out previously specified options from available options
|
|
||||||
prev_opts = [x.split("=")[0] for x in cwords[1 : cword - 1]]
|
|
||||||
options = [(x, v) for (x, v) in options if x not in prev_opts]
|
|
||||||
# filter options by current input
|
|
||||||
options = [(k, v) for k, v in options if k.startswith(current)]
|
|
||||||
# get completion type given cwords and available subcommand options
|
|
||||||
completion_type = get_path_completion_type(
|
|
||||||
cwords,
|
|
||||||
cword,
|
|
||||||
subcommand.parser.option_list_all,
|
|
||||||
)
|
|
||||||
# get completion files and directories if ``completion_type`` is
|
|
||||||
# ``<file>``, ``<dir>`` or ``<path>``
|
|
||||||
if completion_type:
|
|
||||||
paths = auto_complete_paths(current, completion_type)
|
|
||||||
options = [(path, 0) for path in paths]
|
|
||||||
for option in options:
|
|
||||||
opt_label = option[0]
|
|
||||||
# append '=' to options which require args
|
|
||||||
if option[1] and option[0][:2] == "--":
|
|
||||||
opt_label += "="
|
|
||||||
print(opt_label)
|
|
||||||
else:
|
|
||||||
# show main parser options only when necessary
|
|
||||||
|
|
||||||
opts = [i.option_list for i in parser.option_groups]
|
|
||||||
opts.append(parser.option_list)
|
|
||||||
flattened_opts = chain.from_iterable(opts)
|
|
||||||
if current.startswith("-"):
|
|
||||||
for opt in flattened_opts:
|
|
||||||
if opt.help != optparse.SUPPRESS_HELP:
|
|
||||||
subcommands += opt._long_opts + opt._short_opts
|
|
||||||
else:
|
|
||||||
# get completion type given cwords and all available options
|
|
||||||
completion_type = get_path_completion_type(cwords, cword, flattened_opts)
|
|
||||||
if completion_type:
|
|
||||||
subcommands = list(auto_complete_paths(current, completion_type))
|
|
||||||
|
|
||||||
print(" ".join([x for x in subcommands if x.startswith(current)]))
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
|
|
||||||
def get_path_completion_type(
|
|
||||||
cwords: List[str], cword: int, opts: Iterable[Any]
|
|
||||||
) -> Optional[str]:
|
|
||||||
"""Get the type of path completion (``file``, ``dir``, ``path`` or None)
|
|
||||||
|
|
||||||
:param cwords: same as the environmental variable ``COMP_WORDS``
|
|
||||||
:param cword: same as the environmental variable ``COMP_CWORD``
|
|
||||||
:param opts: The available options to check
|
|
||||||
:return: path completion type (``file``, ``dir``, ``path`` or None)
|
|
||||||
"""
|
|
||||||
if cword < 2 or not cwords[cword - 2].startswith("-"):
|
|
||||||
return None
|
|
||||||
for opt in opts:
|
|
||||||
if opt.help == optparse.SUPPRESS_HELP:
|
|
||||||
continue
|
|
||||||
for o in str(opt).split("/"):
|
|
||||||
if cwords[cword - 2].split("=")[0] == o:
|
|
||||||
if not opt.metavar or any(
|
|
||||||
x in ("path", "file", "dir") for x in opt.metavar.split("/")
|
|
||||||
):
|
|
||||||
return opt.metavar
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
def auto_complete_paths(current: str, completion_type: str) -> Iterable[str]:
|
|
||||||
"""If ``completion_type`` is ``file`` or ``path``, list all regular files
|
|
||||||
and directories starting with ``current``; otherwise only list directories
|
|
||||||
starting with ``current``.
|
|
||||||
|
|
||||||
:param current: The word to be completed
|
|
||||||
:param completion_type: path completion type(``file``, ``path`` or ``dir``)
|
|
||||||
:return: A generator of regular files and/or directories
|
|
||||||
"""
|
|
||||||
directory, filename = os.path.split(current)
|
|
||||||
current_path = os.path.abspath(directory)
|
|
||||||
# Don't complete paths if they can't be accessed
|
|
||||||
if not os.access(current_path, os.R_OK):
|
|
||||||
return
|
|
||||||
filename = os.path.normcase(filename)
|
|
||||||
# list all files that start with ``filename``
|
|
||||||
file_list = (
|
|
||||||
x for x in os.listdir(current_path) if os.path.normcase(x).startswith(filename)
|
|
||||||
)
|
|
||||||
for f in file_list:
|
|
||||||
opt = os.path.join(current_path, f)
|
|
||||||
comp_file = os.path.normcase(os.path.join(directory, f))
|
|
||||||
# complete regular files when there is not ``<dir>`` after option
|
|
||||||
# complete directories when there is ``<file>``, ``<path>`` or
|
|
||||||
# ``<dir>``after option
|
|
||||||
if completion_type != "dir" and os.path.isfile(opt):
|
|
||||||
yield comp_file
|
|
||||||
elif os.path.isdir(opt):
|
|
||||||
yield os.path.join(comp_file, "")
|
|
@ -1,236 +0,0 @@
|
|||||||
"""Base Command class, and related routines"""
|
|
||||||
|
|
||||||
import functools
|
|
||||||
import logging
|
|
||||||
import logging.config
|
|
||||||
import optparse
|
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
import traceback
|
|
||||||
from optparse import Values
|
|
||||||
from typing import Any, Callable, List, Optional, Tuple
|
|
||||||
|
|
||||||
from pip._vendor.rich import traceback as rich_traceback
|
|
||||||
|
|
||||||
from pip._internal.cli import cmdoptions
|
|
||||||
from pip._internal.cli.command_context import CommandContextMixIn
|
|
||||||
from pip._internal.cli.parser import ConfigOptionParser, UpdatingDefaultsHelpFormatter
|
|
||||||
from pip._internal.cli.status_codes import (
|
|
||||||
ERROR,
|
|
||||||
PREVIOUS_BUILD_DIR_ERROR,
|
|
||||||
UNKNOWN_ERROR,
|
|
||||||
VIRTUALENV_NOT_FOUND,
|
|
||||||
)
|
|
||||||
from pip._internal.exceptions import (
|
|
||||||
BadCommand,
|
|
||||||
CommandError,
|
|
||||||
DiagnosticPipError,
|
|
||||||
InstallationError,
|
|
||||||
NetworkConnectionError,
|
|
||||||
PreviousBuildDirError,
|
|
||||||
UninstallationError,
|
|
||||||
)
|
|
||||||
from pip._internal.utils.filesystem import check_path_owner
|
|
||||||
from pip._internal.utils.logging import BrokenStdoutLoggingError, setup_logging
|
|
||||||
from pip._internal.utils.misc import get_prog, normalize_path
|
|
||||||
from pip._internal.utils.temp_dir import TempDirectoryTypeRegistry as TempDirRegistry
|
|
||||||
from pip._internal.utils.temp_dir import global_tempdir_manager, tempdir_registry
|
|
||||||
from pip._internal.utils.virtualenv import running_under_virtualenv
|
|
||||||
|
|
||||||
__all__ = ["Command"]
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class Command(CommandContextMixIn):
|
|
||||||
usage: str = ""
|
|
||||||
ignore_require_venv: bool = False
|
|
||||||
|
|
||||||
def __init__(self, name: str, summary: str, isolated: bool = False) -> None:
|
|
||||||
super().__init__()
|
|
||||||
|
|
||||||
self.name = name
|
|
||||||
self.summary = summary
|
|
||||||
self.parser = ConfigOptionParser(
|
|
||||||
usage=self.usage,
|
|
||||||
prog=f"{get_prog()} {name}",
|
|
||||||
formatter=UpdatingDefaultsHelpFormatter(),
|
|
||||||
add_help_option=False,
|
|
||||||
name=name,
|
|
||||||
description=self.__doc__,
|
|
||||||
isolated=isolated,
|
|
||||||
)
|
|
||||||
|
|
||||||
self.tempdir_registry: Optional[TempDirRegistry] = None
|
|
||||||
|
|
||||||
# Commands should add options to this option group
|
|
||||||
optgroup_name = f"{self.name.capitalize()} Options"
|
|
||||||
self.cmd_opts = optparse.OptionGroup(self.parser, optgroup_name)
|
|
||||||
|
|
||||||
# Add the general options
|
|
||||||
gen_opts = cmdoptions.make_option_group(
|
|
||||||
cmdoptions.general_group,
|
|
||||||
self.parser,
|
|
||||||
)
|
|
||||||
self.parser.add_option_group(gen_opts)
|
|
||||||
|
|
||||||
self.add_options()
|
|
||||||
|
|
||||||
def add_options(self) -> None:
|
|
||||||
pass
|
|
||||||
|
|
||||||
def handle_pip_version_check(self, options: Values) -> None:
|
|
||||||
"""
|
|
||||||
This is a no-op so that commands by default do not do the pip version
|
|
||||||
check.
|
|
||||||
"""
|
|
||||||
# Make sure we do the pip version check if the index_group options
|
|
||||||
# are present.
|
|
||||||
assert not hasattr(options, "no_index")
|
|
||||||
|
|
||||||
def run(self, options: Values, args: List[str]) -> int:
|
|
||||||
raise NotImplementedError
|
|
||||||
|
|
||||||
def parse_args(self, args: List[str]) -> Tuple[Values, List[str]]:
|
|
||||||
# factored out for testability
|
|
||||||
return self.parser.parse_args(args)
|
|
||||||
|
|
||||||
def main(self, args: List[str]) -> int:
|
|
||||||
try:
|
|
||||||
with self.main_context():
|
|
||||||
return self._main(args)
|
|
||||||
finally:
|
|
||||||
logging.shutdown()
|
|
||||||
|
|
||||||
def _main(self, args: List[str]) -> int:
|
|
||||||
# We must initialize this before the tempdir manager, otherwise the
|
|
||||||
# configuration would not be accessible by the time we clean up the
|
|
||||||
# tempdir manager.
|
|
||||||
self.tempdir_registry = self.enter_context(tempdir_registry())
|
|
||||||
# Intentionally set as early as possible so globally-managed temporary
|
|
||||||
# directories are available to the rest of the code.
|
|
||||||
self.enter_context(global_tempdir_manager())
|
|
||||||
|
|
||||||
options, args = self.parse_args(args)
|
|
||||||
|
|
||||||
# Set verbosity so that it can be used elsewhere.
|
|
||||||
self.verbosity = options.verbose - options.quiet
|
|
||||||
|
|
||||||
level_number = setup_logging(
|
|
||||||
verbosity=self.verbosity,
|
|
||||||
no_color=options.no_color,
|
|
||||||
user_log_file=options.log,
|
|
||||||
)
|
|
||||||
|
|
||||||
always_enabled_features = set(options.features_enabled) & set(
|
|
||||||
cmdoptions.ALWAYS_ENABLED_FEATURES
|
|
||||||
)
|
|
||||||
if always_enabled_features:
|
|
||||||
logger.warning(
|
|
||||||
"The following features are always enabled: %s. ",
|
|
||||||
", ".join(sorted(always_enabled_features)),
|
|
||||||
)
|
|
||||||
|
|
||||||
# Make sure that the --python argument isn't specified after the
|
|
||||||
# subcommand. We can tell, because if --python was specified,
|
|
||||||
# we should only reach this point if we're running in the created
|
|
||||||
# subprocess, which has the _PIP_RUNNING_IN_SUBPROCESS environment
|
|
||||||
# variable set.
|
|
||||||
if options.python and "_PIP_RUNNING_IN_SUBPROCESS" not in os.environ:
|
|
||||||
logger.critical(
|
|
||||||
"The --python option must be placed before the pip subcommand name"
|
|
||||||
)
|
|
||||||
sys.exit(ERROR)
|
|
||||||
|
|
||||||
# TODO: Try to get these passing down from the command?
|
|
||||||
# without resorting to os.environ to hold these.
|
|
||||||
# This also affects isolated builds and it should.
|
|
||||||
|
|
||||||
if options.no_input:
|
|
||||||
os.environ["PIP_NO_INPUT"] = "1"
|
|
||||||
|
|
||||||
if options.exists_action:
|
|
||||||
os.environ["PIP_EXISTS_ACTION"] = " ".join(options.exists_action)
|
|
||||||
|
|
||||||
if options.require_venv and not self.ignore_require_venv:
|
|
||||||
# If a venv is required check if it can really be found
|
|
||||||
if not running_under_virtualenv():
|
|
||||||
logger.critical("Could not find an activated virtualenv (required).")
|
|
||||||
sys.exit(VIRTUALENV_NOT_FOUND)
|
|
||||||
|
|
||||||
if options.cache_dir:
|
|
||||||
options.cache_dir = normalize_path(options.cache_dir)
|
|
||||||
if not check_path_owner(options.cache_dir):
|
|
||||||
logger.warning(
|
|
||||||
"The directory '%s' or its parent directory is not owned "
|
|
||||||
"or is not writable by the current user. The cache "
|
|
||||||
"has been disabled. Check the permissions and owner of "
|
|
||||||
"that directory. If executing pip with sudo, you should "
|
|
||||||
"use sudo's -H flag.",
|
|
||||||
options.cache_dir,
|
|
||||||
)
|
|
||||||
options.cache_dir = None
|
|
||||||
|
|
||||||
def intercepts_unhandled_exc(
|
|
||||||
run_func: Callable[..., int]
|
|
||||||
) -> Callable[..., int]:
|
|
||||||
@functools.wraps(run_func)
|
|
||||||
def exc_logging_wrapper(*args: Any) -> int:
|
|
||||||
try:
|
|
||||||
status = run_func(*args)
|
|
||||||
assert isinstance(status, int)
|
|
||||||
return status
|
|
||||||
except DiagnosticPipError as exc:
|
|
||||||
logger.error("[present-rich] %s", exc)
|
|
||||||
logger.debug("Exception information:", exc_info=True)
|
|
||||||
|
|
||||||
return ERROR
|
|
||||||
except PreviousBuildDirError as exc:
|
|
||||||
logger.critical(str(exc))
|
|
||||||
logger.debug("Exception information:", exc_info=True)
|
|
||||||
|
|
||||||
return PREVIOUS_BUILD_DIR_ERROR
|
|
||||||
except (
|
|
||||||
InstallationError,
|
|
||||||
UninstallationError,
|
|
||||||
BadCommand,
|
|
||||||
NetworkConnectionError,
|
|
||||||
) as exc:
|
|
||||||
logger.critical(str(exc))
|
|
||||||
logger.debug("Exception information:", exc_info=True)
|
|
||||||
|
|
||||||
return ERROR
|
|
||||||
except CommandError as exc:
|
|
||||||
logger.critical("%s", exc)
|
|
||||||
logger.debug("Exception information:", exc_info=True)
|
|
||||||
|
|
||||||
return ERROR
|
|
||||||
except BrokenStdoutLoggingError:
|
|
||||||
# Bypass our logger and write any remaining messages to
|
|
||||||
# stderr because stdout no longer works.
|
|
||||||
print("ERROR: Pipe to stdout was broken", file=sys.stderr)
|
|
||||||
if level_number <= logging.DEBUG:
|
|
||||||
traceback.print_exc(file=sys.stderr)
|
|
||||||
|
|
||||||
return ERROR
|
|
||||||
except KeyboardInterrupt:
|
|
||||||
logger.critical("Operation cancelled by user")
|
|
||||||
logger.debug("Exception information:", exc_info=True)
|
|
||||||
|
|
||||||
return ERROR
|
|
||||||
except BaseException:
|
|
||||||
logger.critical("Exception:", exc_info=True)
|
|
||||||
|
|
||||||
return UNKNOWN_ERROR
|
|
||||||
|
|
||||||
return exc_logging_wrapper
|
|
||||||
|
|
||||||
try:
|
|
||||||
if not options.debug_mode:
|
|
||||||
run = intercepts_unhandled_exc(self.run)
|
|
||||||
else:
|
|
||||||
run = self.run
|
|
||||||
rich_traceback.install(show_locals=True)
|
|
||||||
return run(options, args)
|
|
||||||
finally:
|
|
||||||
self.handle_pip_version_check(options)
|
|
File diff suppressed because it is too large
Load Diff
@ -1,27 +0,0 @@
|
|||||||
from contextlib import ExitStack, contextmanager
|
|
||||||
from typing import ContextManager, Generator, TypeVar
|
|
||||||
|
|
||||||
_T = TypeVar("_T", covariant=True)
|
|
||||||
|
|
||||||
|
|
||||||
class CommandContextMixIn:
|
|
||||||
def __init__(self) -> None:
|
|
||||||
super().__init__()
|
|
||||||
self._in_main_context = False
|
|
||||||
self._main_context = ExitStack()
|
|
||||||
|
|
||||||
@contextmanager
|
|
||||||
def main_context(self) -> Generator[None, None, None]:
|
|
||||||
assert not self._in_main_context
|
|
||||||
|
|
||||||
self._in_main_context = True
|
|
||||||
try:
|
|
||||||
with self._main_context:
|
|
||||||
yield
|
|
||||||
finally:
|
|
||||||
self._in_main_context = False
|
|
||||||
|
|
||||||
def enter_context(self, context_provider: ContextManager[_T]) -> _T:
|
|
||||||
assert self._in_main_context
|
|
||||||
|
|
||||||
return self._main_context.enter_context(context_provider)
|
|
@ -1,79 +0,0 @@
|
|||||||
"""Primary application entrypoint.
|
|
||||||
"""
|
|
||||||
import locale
|
|
||||||
import logging
|
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
import warnings
|
|
||||||
from typing import List, Optional
|
|
||||||
|
|
||||||
from pip._internal.cli.autocompletion import autocomplete
|
|
||||||
from pip._internal.cli.main_parser import parse_command
|
|
||||||
from pip._internal.commands import create_command
|
|
||||||
from pip._internal.exceptions import PipError
|
|
||||||
from pip._internal.utils import deprecation
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
# Do not import and use main() directly! Using it directly is actively
|
|
||||||
# discouraged by pip's maintainers. The name, location and behavior of
|
|
||||||
# this function is subject to change, so calling it directly is not
|
|
||||||
# portable across different pip versions.
|
|
||||||
|
|
||||||
# In addition, running pip in-process is unsupported and unsafe. This is
|
|
||||||
# elaborated in detail at
|
|
||||||
# https://pip.pypa.io/en/stable/user_guide/#using-pip-from-your-program.
|
|
||||||
# That document also provides suggestions that should work for nearly
|
|
||||||
# all users that are considering importing and using main() directly.
|
|
||||||
|
|
||||||
# However, we know that certain users will still want to invoke pip
|
|
||||||
# in-process. If you understand and accept the implications of using pip
|
|
||||||
# in an unsupported manner, the best approach is to use runpy to avoid
|
|
||||||
# depending on the exact location of this entry point.
|
|
||||||
|
|
||||||
# The following example shows how to use runpy to invoke pip in that
|
|
||||||
# case:
|
|
||||||
#
|
|
||||||
# sys.argv = ["pip", your, args, here]
|
|
||||||
# runpy.run_module("pip", run_name="__main__")
|
|
||||||
#
|
|
||||||
# Note that this will exit the process after running, unlike a direct
|
|
||||||
# call to main. As it is not safe to do any processing after calling
|
|
||||||
# main, this should not be an issue in practice.
|
|
||||||
|
|
||||||
|
|
||||||
def main(args: Optional[List[str]] = None) -> int:
|
|
||||||
if args is None:
|
|
||||||
args = sys.argv[1:]
|
|
||||||
|
|
||||||
# Suppress the pkg_resources deprecation warning
|
|
||||||
# Note - we use a module of .*pkg_resources to cover
|
|
||||||
# the normal case (pip._vendor.pkg_resources) and the
|
|
||||||
# devendored case (a bare pkg_resources)
|
|
||||||
warnings.filterwarnings(
|
|
||||||
action="ignore", category=DeprecationWarning, module=".*pkg_resources"
|
|
||||||
)
|
|
||||||
|
|
||||||
# Configure our deprecation warnings to be sent through loggers
|
|
||||||
deprecation.install_warning_logger()
|
|
||||||
|
|
||||||
autocomplete()
|
|
||||||
|
|
||||||
try:
|
|
||||||
cmd_name, cmd_args = parse_command(args)
|
|
||||||
except PipError as exc:
|
|
||||||
sys.stderr.write(f"ERROR: {exc}")
|
|
||||||
sys.stderr.write(os.linesep)
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
# Needed for locale.getpreferredencoding(False) to work
|
|
||||||
# in pip._internal.utils.encoding.auto_decode
|
|
||||||
try:
|
|
||||||
locale.setlocale(locale.LC_ALL, "")
|
|
||||||
except locale.Error as e:
|
|
||||||
# setlocale can apparently crash if locale are uninitialized
|
|
||||||
logger.debug("Ignoring error %s when setting locale", e)
|
|
||||||
command = create_command(cmd_name, isolated=("--isolated" in cmd_args))
|
|
||||||
|
|
||||||
return command.main(cmd_args)
|
|
@ -1,134 +0,0 @@
|
|||||||
"""A single place for constructing and exposing the main parser
|
|
||||||
"""
|
|
||||||
|
|
||||||
import os
|
|
||||||
import subprocess
|
|
||||||
import sys
|
|
||||||
from typing import List, Optional, Tuple
|
|
||||||
|
|
||||||
from pip._internal.build_env import get_runnable_pip
|
|
||||||
from pip._internal.cli import cmdoptions
|
|
||||||
from pip._internal.cli.parser import ConfigOptionParser, UpdatingDefaultsHelpFormatter
|
|
||||||
from pip._internal.commands import commands_dict, get_similar_commands
|
|
||||||
from pip._internal.exceptions import CommandError
|
|
||||||
from pip._internal.utils.misc import get_pip_version, get_prog
|
|
||||||
|
|
||||||
__all__ = ["create_main_parser", "parse_command"]
|
|
||||||
|
|
||||||
|
|
||||||
def create_main_parser() -> ConfigOptionParser:
|
|
||||||
"""Creates and returns the main parser for pip's CLI"""
|
|
||||||
|
|
||||||
parser = ConfigOptionParser(
|
|
||||||
usage="\n%prog <command> [options]",
|
|
||||||
add_help_option=False,
|
|
||||||
formatter=UpdatingDefaultsHelpFormatter(),
|
|
||||||
name="global",
|
|
||||||
prog=get_prog(),
|
|
||||||
)
|
|
||||||
parser.disable_interspersed_args()
|
|
||||||
|
|
||||||
parser.version = get_pip_version()
|
|
||||||
|
|
||||||
# add the general options
|
|
||||||
gen_opts = cmdoptions.make_option_group(cmdoptions.general_group, parser)
|
|
||||||
parser.add_option_group(gen_opts)
|
|
||||||
|
|
||||||
# so the help formatter knows
|
|
||||||
parser.main = True # type: ignore
|
|
||||||
|
|
||||||
# create command listing for description
|
|
||||||
description = [""] + [
|
|
||||||
f"{name:27} {command_info.summary}"
|
|
||||||
for name, command_info in commands_dict.items()
|
|
||||||
]
|
|
||||||
parser.description = "\n".join(description)
|
|
||||||
|
|
||||||
return parser
|
|
||||||
|
|
||||||
|
|
||||||
def identify_python_interpreter(python: str) -> Optional[str]:
|
|
||||||
# If the named file exists, use it.
|
|
||||||
# If it's a directory, assume it's a virtual environment and
|
|
||||||
# look for the environment's Python executable.
|
|
||||||
if os.path.exists(python):
|
|
||||||
if os.path.isdir(python):
|
|
||||||
# bin/python for Unix, Scripts/python.exe for Windows
|
|
||||||
# Try both in case of odd cases like cygwin.
|
|
||||||
for exe in ("bin/python", "Scripts/python.exe"):
|
|
||||||
py = os.path.join(python, exe)
|
|
||||||
if os.path.exists(py):
|
|
||||||
return py
|
|
||||||
else:
|
|
||||||
return python
|
|
||||||
|
|
||||||
# Could not find the interpreter specified
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
def parse_command(args: List[str]) -> Tuple[str, List[str]]:
|
|
||||||
parser = create_main_parser()
|
|
||||||
|
|
||||||
# Note: parser calls disable_interspersed_args(), so the result of this
|
|
||||||
# call is to split the initial args into the general options before the
|
|
||||||
# subcommand and everything else.
|
|
||||||
# For example:
|
|
||||||
# args: ['--timeout=5', 'install', '--user', 'INITools']
|
|
||||||
# general_options: ['--timeout==5']
|
|
||||||
# args_else: ['install', '--user', 'INITools']
|
|
||||||
general_options, args_else = parser.parse_args(args)
|
|
||||||
|
|
||||||
# --python
|
|
||||||
if general_options.python and "_PIP_RUNNING_IN_SUBPROCESS" not in os.environ:
|
|
||||||
# Re-invoke pip using the specified Python interpreter
|
|
||||||
interpreter = identify_python_interpreter(general_options.python)
|
|
||||||
if interpreter is None:
|
|
||||||
raise CommandError(
|
|
||||||
f"Could not locate Python interpreter {general_options.python}"
|
|
||||||
)
|
|
||||||
|
|
||||||
pip_cmd = [
|
|
||||||
interpreter,
|
|
||||||
get_runnable_pip(),
|
|
||||||
]
|
|
||||||
pip_cmd.extend(args)
|
|
||||||
|
|
||||||
# Set a flag so the child doesn't re-invoke itself, causing
|
|
||||||
# an infinite loop.
|
|
||||||
os.environ["_PIP_RUNNING_IN_SUBPROCESS"] = "1"
|
|
||||||
returncode = 0
|
|
||||||
try:
|
|
||||||
proc = subprocess.run(pip_cmd)
|
|
||||||
returncode = proc.returncode
|
|
||||||
except (subprocess.SubprocessError, OSError) as exc:
|
|
||||||
raise CommandError(f"Failed to run pip under {interpreter}: {exc}")
|
|
||||||
sys.exit(returncode)
|
|
||||||
|
|
||||||
# --version
|
|
||||||
if general_options.version:
|
|
||||||
sys.stdout.write(parser.version)
|
|
||||||
sys.stdout.write(os.linesep)
|
|
||||||
sys.exit()
|
|
||||||
|
|
||||||
# pip || pip help -> print_help()
|
|
||||||
if not args_else or (args_else[0] == "help" and len(args_else) == 1):
|
|
||||||
parser.print_help()
|
|
||||||
sys.exit()
|
|
||||||
|
|
||||||
# the subcommand name
|
|
||||||
cmd_name = args_else[0]
|
|
||||||
|
|
||||||
if cmd_name not in commands_dict:
|
|
||||||
guess = get_similar_commands(cmd_name)
|
|
||||||
|
|
||||||
msg = [f'unknown command "{cmd_name}"']
|
|
||||||
if guess:
|
|
||||||
msg.append(f'maybe you meant "{guess}"')
|
|
||||||
|
|
||||||
raise CommandError(" - ".join(msg))
|
|
||||||
|
|
||||||
# all the args without the subcommand
|
|
||||||
cmd_args = args[:]
|
|
||||||
cmd_args.remove(cmd_name)
|
|
||||||
|
|
||||||
return cmd_name, cmd_args
|
|
@ -1,294 +0,0 @@
|
|||||||
"""Base option parser setup"""
|
|
||||||
|
|
||||||
import logging
|
|
||||||
import optparse
|
|
||||||
import shutil
|
|
||||||
import sys
|
|
||||||
import textwrap
|
|
||||||
from contextlib import suppress
|
|
||||||
from typing import Any, Dict, Generator, List, Tuple
|
|
||||||
|
|
||||||
from pip._internal.cli.status_codes import UNKNOWN_ERROR
|
|
||||||
from pip._internal.configuration import Configuration, ConfigurationError
|
|
||||||
from pip._internal.utils.misc import redact_auth_from_url, strtobool
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class PrettyHelpFormatter(optparse.IndentedHelpFormatter):
|
|
||||||
"""A prettier/less verbose help formatter for optparse."""
|
|
||||||
|
|
||||||
def __init__(self, *args: Any, **kwargs: Any) -> None:
|
|
||||||
# help position must be aligned with __init__.parseopts.description
|
|
||||||
kwargs["max_help_position"] = 30
|
|
||||||
kwargs["indent_increment"] = 1
|
|
||||||
kwargs["width"] = shutil.get_terminal_size()[0] - 2
|
|
||||||
super().__init__(*args, **kwargs)
|
|
||||||
|
|
||||||
def format_option_strings(self, option: optparse.Option) -> str:
|
|
||||||
return self._format_option_strings(option)
|
|
||||||
|
|
||||||
def _format_option_strings(
|
|
||||||
self, option: optparse.Option, mvarfmt: str = " <{}>", optsep: str = ", "
|
|
||||||
) -> str:
|
|
||||||
"""
|
|
||||||
Return a comma-separated list of option strings and metavars.
|
|
||||||
|
|
||||||
:param option: tuple of (short opt, long opt), e.g: ('-f', '--format')
|
|
||||||
:param mvarfmt: metavar format string
|
|
||||||
:param optsep: separator
|
|
||||||
"""
|
|
||||||
opts = []
|
|
||||||
|
|
||||||
if option._short_opts:
|
|
||||||
opts.append(option._short_opts[0])
|
|
||||||
if option._long_opts:
|
|
||||||
opts.append(option._long_opts[0])
|
|
||||||
if len(opts) > 1:
|
|
||||||
opts.insert(1, optsep)
|
|
||||||
|
|
||||||
if option.takes_value():
|
|
||||||
assert option.dest is not None
|
|
||||||
metavar = option.metavar or option.dest.lower()
|
|
||||||
opts.append(mvarfmt.format(metavar.lower()))
|
|
||||||
|
|
||||||
return "".join(opts)
|
|
||||||
|
|
||||||
def format_heading(self, heading: str) -> str:
|
|
||||||
if heading == "Options":
|
|
||||||
return ""
|
|
||||||
return heading + ":\n"
|
|
||||||
|
|
||||||
def format_usage(self, usage: str) -> str:
|
|
||||||
"""
|
|
||||||
Ensure there is only one newline between usage and the first heading
|
|
||||||
if there is no description.
|
|
||||||
"""
|
|
||||||
msg = "\nUsage: {}\n".format(self.indent_lines(textwrap.dedent(usage), " "))
|
|
||||||
return msg
|
|
||||||
|
|
||||||
def format_description(self, description: str) -> str:
|
|
||||||
# leave full control over description to us
|
|
||||||
if description:
|
|
||||||
if hasattr(self.parser, "main"):
|
|
||||||
label = "Commands"
|
|
||||||
else:
|
|
||||||
label = "Description"
|
|
||||||
# some doc strings have initial newlines, some don't
|
|
||||||
description = description.lstrip("\n")
|
|
||||||
# some doc strings have final newlines and spaces, some don't
|
|
||||||
description = description.rstrip()
|
|
||||||
# dedent, then reindent
|
|
||||||
description = self.indent_lines(textwrap.dedent(description), " ")
|
|
||||||
description = f"{label}:\n{description}\n"
|
|
||||||
return description
|
|
||||||
else:
|
|
||||||
return ""
|
|
||||||
|
|
||||||
def format_epilog(self, epilog: str) -> str:
|
|
||||||
# leave full control over epilog to us
|
|
||||||
if epilog:
|
|
||||||
return epilog
|
|
||||||
else:
|
|
||||||
return ""
|
|
||||||
|
|
||||||
def indent_lines(self, text: str, indent: str) -> str:
|
|
||||||
new_lines = [indent + line for line in text.split("\n")]
|
|
||||||
return "\n".join(new_lines)
|
|
||||||
|
|
||||||
|
|
||||||
class UpdatingDefaultsHelpFormatter(PrettyHelpFormatter):
|
|
||||||
"""Custom help formatter for use in ConfigOptionParser.
|
|
||||||
|
|
||||||
This is updates the defaults before expanding them, allowing
|
|
||||||
them to show up correctly in the help listing.
|
|
||||||
|
|
||||||
Also redact auth from url type options
|
|
||||||
"""
|
|
||||||
|
|
||||||
def expand_default(self, option: optparse.Option) -> str:
|
|
||||||
default_values = None
|
|
||||||
if self.parser is not None:
|
|
||||||
assert isinstance(self.parser, ConfigOptionParser)
|
|
||||||
self.parser._update_defaults(self.parser.defaults)
|
|
||||||
assert option.dest is not None
|
|
||||||
default_values = self.parser.defaults.get(option.dest)
|
|
||||||
help_text = super().expand_default(option)
|
|
||||||
|
|
||||||
if default_values and option.metavar == "URL":
|
|
||||||
if isinstance(default_values, str):
|
|
||||||
default_values = [default_values]
|
|
||||||
|
|
||||||
# If its not a list, we should abort and just return the help text
|
|
||||||
if not isinstance(default_values, list):
|
|
||||||
default_values = []
|
|
||||||
|
|
||||||
for val in default_values:
|
|
||||||
help_text = help_text.replace(val, redact_auth_from_url(val))
|
|
||||||
|
|
||||||
return help_text
|
|
||||||
|
|
||||||
|
|
||||||
class CustomOptionParser(optparse.OptionParser):
|
|
||||||
def insert_option_group(
|
|
||||||
self, idx: int, *args: Any, **kwargs: Any
|
|
||||||
) -> optparse.OptionGroup:
|
|
||||||
"""Insert an OptionGroup at a given position."""
|
|
||||||
group = self.add_option_group(*args, **kwargs)
|
|
||||||
|
|
||||||
self.option_groups.pop()
|
|
||||||
self.option_groups.insert(idx, group)
|
|
||||||
|
|
||||||
return group
|
|
||||||
|
|
||||||
@property
|
|
||||||
def option_list_all(self) -> List[optparse.Option]:
|
|
||||||
"""Get a list of all options, including those in option groups."""
|
|
||||||
res = self.option_list[:]
|
|
||||||
for i in self.option_groups:
|
|
||||||
res.extend(i.option_list)
|
|
||||||
|
|
||||||
return res
|
|
||||||
|
|
||||||
|
|
||||||
class ConfigOptionParser(CustomOptionParser):
|
|
||||||
"""Custom option parser which updates its defaults by checking the
|
|
||||||
configuration files and environmental variables"""
|
|
||||||
|
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
*args: Any,
|
|
||||||
name: str,
|
|
||||||
isolated: bool = False,
|
|
||||||
**kwargs: Any,
|
|
||||||
) -> None:
|
|
||||||
self.name = name
|
|
||||||
self.config = Configuration(isolated)
|
|
||||||
|
|
||||||
assert self.name
|
|
||||||
super().__init__(*args, **kwargs)
|
|
||||||
|
|
||||||
def check_default(self, option: optparse.Option, key: str, val: Any) -> Any:
|
|
||||||
try:
|
|
||||||
return option.check_value(key, val)
|
|
||||||
except optparse.OptionValueError as exc:
|
|
||||||
print(f"An error occurred during configuration: {exc}")
|
|
||||||
sys.exit(3)
|
|
||||||
|
|
||||||
def _get_ordered_configuration_items(
|
|
||||||
self,
|
|
||||||
) -> Generator[Tuple[str, Any], None, None]:
|
|
||||||
# Configuration gives keys in an unordered manner. Order them.
|
|
||||||
override_order = ["global", self.name, ":env:"]
|
|
||||||
|
|
||||||
# Pool the options into different groups
|
|
||||||
section_items: Dict[str, List[Tuple[str, Any]]] = {
|
|
||||||
name: [] for name in override_order
|
|
||||||
}
|
|
||||||
for section_key, val in self.config.items():
|
|
||||||
# ignore empty values
|
|
||||||
if not val:
|
|
||||||
logger.debug(
|
|
||||||
"Ignoring configuration key '%s' as it's value is empty.",
|
|
||||||
section_key,
|
|
||||||
)
|
|
||||||
continue
|
|
||||||
|
|
||||||
section, key = section_key.split(".", 1)
|
|
||||||
if section in override_order:
|
|
||||||
section_items[section].append((key, val))
|
|
||||||
|
|
||||||
# Yield each group in their override order
|
|
||||||
for section in override_order:
|
|
||||||
for key, val in section_items[section]:
|
|
||||||
yield key, val
|
|
||||||
|
|
||||||
def _update_defaults(self, defaults: Dict[str, Any]) -> Dict[str, Any]:
|
|
||||||
"""Updates the given defaults with values from the config files and
|
|
||||||
the environ. Does a little special handling for certain types of
|
|
||||||
options (lists)."""
|
|
||||||
|
|
||||||
# Accumulate complex default state.
|
|
||||||
self.values = optparse.Values(self.defaults)
|
|
||||||
late_eval = set()
|
|
||||||
# Then set the options with those values
|
|
||||||
for key, val in self._get_ordered_configuration_items():
|
|
||||||
# '--' because configuration supports only long names
|
|
||||||
option = self.get_option("--" + key)
|
|
||||||
|
|
||||||
# Ignore options not present in this parser. E.g. non-globals put
|
|
||||||
# in [global] by users that want them to apply to all applicable
|
|
||||||
# commands.
|
|
||||||
if option is None:
|
|
||||||
continue
|
|
||||||
|
|
||||||
assert option.dest is not None
|
|
||||||
|
|
||||||
if option.action in ("store_true", "store_false"):
|
|
||||||
try:
|
|
||||||
val = strtobool(val)
|
|
||||||
except ValueError:
|
|
||||||
self.error(
|
|
||||||
"{} is not a valid value for {} option, " # noqa
|
|
||||||
"please specify a boolean value like yes/no, "
|
|
||||||
"true/false or 1/0 instead.".format(val, key)
|
|
||||||
)
|
|
||||||
elif option.action == "count":
|
|
||||||
with suppress(ValueError):
|
|
||||||
val = strtobool(val)
|
|
||||||
with suppress(ValueError):
|
|
||||||
val = int(val)
|
|
||||||
if not isinstance(val, int) or val < 0:
|
|
||||||
self.error(
|
|
||||||
"{} is not a valid value for {} option, " # noqa
|
|
||||||
"please instead specify either a non-negative integer "
|
|
||||||
"or a boolean value like yes/no or false/true "
|
|
||||||
"which is equivalent to 1/0.".format(val, key)
|
|
||||||
)
|
|
||||||
elif option.action == "append":
|
|
||||||
val = val.split()
|
|
||||||
val = [self.check_default(option, key, v) for v in val]
|
|
||||||
elif option.action == "callback":
|
|
||||||
assert option.callback is not None
|
|
||||||
late_eval.add(option.dest)
|
|
||||||
opt_str = option.get_opt_string()
|
|
||||||
val = option.convert_value(opt_str, val)
|
|
||||||
# From take_action
|
|
||||||
args = option.callback_args or ()
|
|
||||||
kwargs = option.callback_kwargs or {}
|
|
||||||
option.callback(option, opt_str, val, self, *args, **kwargs)
|
|
||||||
else:
|
|
||||||
val = self.check_default(option, key, val)
|
|
||||||
|
|
||||||
defaults[option.dest] = val
|
|
||||||
|
|
||||||
for key in late_eval:
|
|
||||||
defaults[key] = getattr(self.values, key)
|
|
||||||
self.values = None
|
|
||||||
return defaults
|
|
||||||
|
|
||||||
def get_default_values(self) -> optparse.Values:
|
|
||||||
"""Overriding to make updating the defaults after instantiation of
|
|
||||||
the option parser possible, _update_defaults() does the dirty work."""
|
|
||||||
if not self.process_default_values:
|
|
||||||
# Old, pre-Optik 1.5 behaviour.
|
|
||||||
return optparse.Values(self.defaults)
|
|
||||||
|
|
||||||
# Load the configuration, or error out in case of an error
|
|
||||||
try:
|
|
||||||
self.config.load()
|
|
||||||
except ConfigurationError as err:
|
|
||||||
self.exit(UNKNOWN_ERROR, str(err))
|
|
||||||
|
|
||||||
defaults = self._update_defaults(self.defaults.copy()) # ours
|
|
||||||
for option in self._get_all_options():
|
|
||||||
assert option.dest is not None
|
|
||||||
default = defaults.get(option.dest)
|
|
||||||
if isinstance(default, str):
|
|
||||||
opt_str = option.get_opt_string()
|
|
||||||
defaults[option.dest] = option.check_value(opt_str, default)
|
|
||||||
return optparse.Values(defaults)
|
|
||||||
|
|
||||||
def error(self, msg: str) -> None:
|
|
||||||
self.print_usage(sys.stderr)
|
|
||||||
self.exit(UNKNOWN_ERROR, f"{msg}\n")
|
|
@ -1,68 +0,0 @@
|
|||||||
import functools
|
|
||||||
from typing import Callable, Generator, Iterable, Iterator, Optional, Tuple
|
|
||||||
|
|
||||||
from pip._vendor.rich.progress import (
|
|
||||||
BarColumn,
|
|
||||||
DownloadColumn,
|
|
||||||
FileSizeColumn,
|
|
||||||
Progress,
|
|
||||||
ProgressColumn,
|
|
||||||
SpinnerColumn,
|
|
||||||
TextColumn,
|
|
||||||
TimeElapsedColumn,
|
|
||||||
TimeRemainingColumn,
|
|
||||||
TransferSpeedColumn,
|
|
||||||
)
|
|
||||||
|
|
||||||
from pip._internal.utils.logging import get_indentation
|
|
||||||
|
|
||||||
DownloadProgressRenderer = Callable[[Iterable[bytes]], Iterator[bytes]]
|
|
||||||
|
|
||||||
|
|
||||||
def _rich_progress_bar(
|
|
||||||
iterable: Iterable[bytes],
|
|
||||||
*,
|
|
||||||
bar_type: str,
|
|
||||||
size: int,
|
|
||||||
) -> Generator[bytes, None, None]:
|
|
||||||
assert bar_type == "on", "This should only be used in the default mode."
|
|
||||||
|
|
||||||
if not size:
|
|
||||||
total = float("inf")
|
|
||||||
columns: Tuple[ProgressColumn, ...] = (
|
|
||||||
TextColumn("[progress.description]{task.description}"),
|
|
||||||
SpinnerColumn("line", speed=1.5),
|
|
||||||
FileSizeColumn(),
|
|
||||||
TransferSpeedColumn(),
|
|
||||||
TimeElapsedColumn(),
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
total = size
|
|
||||||
columns = (
|
|
||||||
TextColumn("[progress.description]{task.description}"),
|
|
||||||
BarColumn(),
|
|
||||||
DownloadColumn(),
|
|
||||||
TransferSpeedColumn(),
|
|
||||||
TextColumn("eta"),
|
|
||||||
TimeRemainingColumn(),
|
|
||||||
)
|
|
||||||
|
|
||||||
progress = Progress(*columns, refresh_per_second=30)
|
|
||||||
task_id = progress.add_task(" " * (get_indentation() + 2), total=total)
|
|
||||||
with progress:
|
|
||||||
for chunk in iterable:
|
|
||||||
yield chunk
|
|
||||||
progress.update(task_id, advance=len(chunk))
|
|
||||||
|
|
||||||
|
|
||||||
def get_download_progress_renderer(
|
|
||||||
*, bar_type: str, size: Optional[int] = None
|
|
||||||
) -> DownloadProgressRenderer:
|
|
||||||
"""Get an object that can be used to render the download progress.
|
|
||||||
|
|
||||||
Returns a callable, that takes an iterable to "wrap".
|
|
||||||
"""
|
|
||||||
if bar_type == "on":
|
|
||||||
return functools.partial(_rich_progress_bar, bar_type=bar_type, size=size)
|
|
||||||
else:
|
|
||||||
return iter # no-op, when passed an iterator
|
|
File diff suppressed because it is too large
Load Diff
@ -1,159 +0,0 @@
|
|||||||
import contextlib
|
|
||||||
import itertools
|
|
||||||
import logging
|
|
||||||
import sys
|
|
||||||
import time
|
|
||||||
from typing import IO, Generator, Optional
|
|
||||||
|
|
||||||
from pip._internal.utils.compat import WINDOWS
|
|
||||||
from pip._internal.utils.logging import get_indentation
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class SpinnerInterface:
|
|
||||||
def spin(self) -> None:
|
|
||||||
raise NotImplementedError()
|
|
||||||
|
|
||||||
def finish(self, final_status: str) -> None:
|
|
||||||
raise NotImplementedError()
|
|
||||||
|
|
||||||
|
|
||||||
class InteractiveSpinner(SpinnerInterface):
|
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
message: str,
|
|
||||||
file: Optional[IO[str]] = None,
|
|
||||||
spin_chars: str = "-\\|/",
|
|
||||||
# Empirically, 8 updates/second looks nice
|
|
||||||
min_update_interval_seconds: float = 0.125,
|
|
||||||
):
|
|
||||||
self._message = message
|
|
||||||
if file is None:
|
|
||||||
file = sys.stdout
|
|
||||||
self._file = file
|
|
||||||
self._rate_limiter = RateLimiter(min_update_interval_seconds)
|
|
||||||
self._finished = False
|
|
||||||
|
|
||||||
self._spin_cycle = itertools.cycle(spin_chars)
|
|
||||||
|
|
||||||
self._file.write(" " * get_indentation() + self._message + " ... ")
|
|
||||||
self._width = 0
|
|
||||||
|
|
||||||
def _write(self, status: str) -> None:
|
|
||||||
assert not self._finished
|
|
||||||
# Erase what we wrote before by backspacing to the beginning, writing
|
|
||||||
# spaces to overwrite the old text, and then backspacing again
|
|
||||||
backup = "\b" * self._width
|
|
||||||
self._file.write(backup + " " * self._width + backup)
|
|
||||||
# Now we have a blank slate to add our status
|
|
||||||
self._file.write(status)
|
|
||||||
self._width = len(status)
|
|
||||||
self._file.flush()
|
|
||||||
self._rate_limiter.reset()
|
|
||||||
|
|
||||||
def spin(self) -> None:
|
|
||||||
if self._finished:
|
|
||||||
return
|
|
||||||
if not self._rate_limiter.ready():
|
|
||||||
return
|
|
||||||
self._write(next(self._spin_cycle))
|
|
||||||
|
|
||||||
def finish(self, final_status: str) -> None:
|
|
||||||
if self._finished:
|
|
||||||
return
|
|
||||||
self._write(final_status)
|
|
||||||
self._file.write("\n")
|
|
||||||
self._file.flush()
|
|
||||||
self._finished = True
|
|
||||||
|
|
||||||
|
|
||||||
# Used for dumb terminals, non-interactive installs (no tty), etc.
|
|
||||||
# We still print updates occasionally (once every 60 seconds by default) to
|
|
||||||
# act as a keep-alive for systems like Travis-CI that take lack-of-output as
|
|
||||||
# an indication that a task has frozen.
|
|
||||||
class NonInteractiveSpinner(SpinnerInterface):
|
|
||||||
def __init__(self, message: str, min_update_interval_seconds: float = 60.0) -> None:
|
|
||||||
self._message = message
|
|
||||||
self._finished = False
|
|
||||||
self._rate_limiter = RateLimiter(min_update_interval_seconds)
|
|
||||||
self._update("started")
|
|
||||||
|
|
||||||
def _update(self, status: str) -> None:
|
|
||||||
assert not self._finished
|
|
||||||
self._rate_limiter.reset()
|
|
||||||
logger.info("%s: %s", self._message, status)
|
|
||||||
|
|
||||||
def spin(self) -> None:
|
|
||||||
if self._finished:
|
|
||||||
return
|
|
||||||
if not self._rate_limiter.ready():
|
|
||||||
return
|
|
||||||
self._update("still running...")
|
|
||||||
|
|
||||||
def finish(self, final_status: str) -> None:
|
|
||||||
if self._finished:
|
|
||||||
return
|
|
||||||
self._update(f"finished with status '{final_status}'")
|
|
||||||
self._finished = True
|
|
||||||
|
|
||||||
|
|
||||||
class RateLimiter:
|
|
||||||
def __init__(self, min_update_interval_seconds: float) -> None:
|
|
||||||
self._min_update_interval_seconds = min_update_interval_seconds
|
|
||||||
self._last_update: float = 0
|
|
||||||
|
|
||||||
def ready(self) -> bool:
|
|
||||||
now = time.time()
|
|
||||||
delta = now - self._last_update
|
|
||||||
return delta >= self._min_update_interval_seconds
|
|
||||||
|
|
||||||
def reset(self) -> None:
|
|
||||||
self._last_update = time.time()
|
|
||||||
|
|
||||||
|
|
||||||
@contextlib.contextmanager
|
|
||||||
def open_spinner(message: str) -> Generator[SpinnerInterface, None, None]:
|
|
||||||
# Interactive spinner goes directly to sys.stdout rather than being routed
|
|
||||||
# through the logging system, but it acts like it has level INFO,
|
|
||||||
# i.e. it's only displayed if we're at level INFO or better.
|
|
||||||
# Non-interactive spinner goes through the logging system, so it is always
|
|
||||||
# in sync with logging configuration.
|
|
||||||
if sys.stdout.isatty() and logger.getEffectiveLevel() <= logging.INFO:
|
|
||||||
spinner: SpinnerInterface = InteractiveSpinner(message)
|
|
||||||
else:
|
|
||||||
spinner = NonInteractiveSpinner(message)
|
|
||||||
try:
|
|
||||||
with hidden_cursor(sys.stdout):
|
|
||||||
yield spinner
|
|
||||||
except KeyboardInterrupt:
|
|
||||||
spinner.finish("canceled")
|
|
||||||
raise
|
|
||||||
except Exception:
|
|
||||||
spinner.finish("error")
|
|
||||||
raise
|
|
||||||
else:
|
|
||||||
spinner.finish("done")
|
|
||||||
|
|
||||||
|
|
||||||
HIDE_CURSOR = "\x1b[?25l"
|
|
||||||
SHOW_CURSOR = "\x1b[?25h"
|
|
||||||
|
|
||||||
|
|
||||||
@contextlib.contextmanager
|
|
||||||
def hidden_cursor(file: IO[str]) -> Generator[None, None, None]:
|
|
||||||
# The Windows terminal does not support the hide/show cursor ANSI codes,
|
|
||||||
# even via colorama. So don't even try.
|
|
||||||
if WINDOWS:
|
|
||||||
yield
|
|
||||||
# We don't want to clutter the output with control characters if we're
|
|
||||||
# writing to a file, or if the user is running with --quiet.
|
|
||||||
# See https://github.com/pypa/pip/issues/3418
|
|
||||||
elif not file.isatty() or logger.getEffectiveLevel() > logging.INFO:
|
|
||||||
yield
|
|
||||||
else:
|
|
||||||
file.write(HIDE_CURSOR)
|
|
||||||
try:
|
|
||||||
yield
|
|
||||||
finally:
|
|
||||||
file.write(SHOW_CURSOR)
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue
Block a user