Mercurial > repos > guerler > springsuite
comparison planemo/lib/python3.7/site-packages/pip/_internal/operations/prepare.py @ 1:56ad4e20f292 draft
"planemo upload commit 6eee67778febed82ddd413c3ca40b3183a3898f1"
| author | guerler |
|---|---|
| date | Fri, 31 Jul 2020 00:32:28 -0400 |
| parents | |
| children |
comparison
equal
deleted
inserted
replaced
| 0:d30785e31577 | 1:56ad4e20f292 |
|---|---|
| 1 """Prepares a distribution for installation | |
| 2 """ | |
| 3 | |
| 4 import logging | |
| 5 import os | |
| 6 | |
| 7 from pip._vendor import requests | |
| 8 | |
| 9 from pip._internal.distributions import ( | |
| 10 make_distribution_for_install_requirement, | |
| 11 ) | |
| 12 from pip._internal.distributions.installed import InstalledDistribution | |
| 13 from pip._internal.download import ( | |
| 14 is_dir_url, is_file_url, is_vcs_url, unpack_url, url_to_path, | |
| 15 ) | |
| 16 from pip._internal.exceptions import ( | |
| 17 DirectoryUrlHashUnsupported, HashUnpinned, InstallationError, | |
| 18 PreviousBuildDirError, VcsHashUnsupported, | |
| 19 ) | |
| 20 from pip._internal.utils.compat import expanduser | |
| 21 from pip._internal.utils.hashes import MissingHashes | |
| 22 from pip._internal.utils.logging import indent_log | |
| 23 from pip._internal.utils.misc import display_path, normalize_path | |
| 24 from pip._internal.utils.typing import MYPY_CHECK_RUNNING | |
| 25 | |
| 26 if MYPY_CHECK_RUNNING: | |
| 27 from typing import Optional | |
| 28 | |
| 29 from pip._internal.distributions import AbstractDistribution | |
| 30 from pip._internal.download import PipSession | |
| 31 from pip._internal.index import PackageFinder | |
| 32 from pip._internal.req.req_install import InstallRequirement | |
| 33 from pip._internal.req.req_tracker import RequirementTracker | |
| 34 | |
| 35 logger = logging.getLogger(__name__) | |
| 36 | |
| 37 | |
| 38 class RequirementPreparer(object): | |
| 39 """Prepares a Requirement | |
| 40 """ | |
| 41 | |
| 42 def __init__( | |
| 43 self, | |
| 44 build_dir, # type: str | |
| 45 download_dir, # type: Optional[str] | |
| 46 src_dir, # type: str | |
| 47 wheel_download_dir, # type: Optional[str] | |
| 48 progress_bar, # type: str | |
| 49 build_isolation, # type: bool | |
| 50 req_tracker # type: RequirementTracker | |
| 51 ): | |
| 52 # type: (...) -> None | |
| 53 super(RequirementPreparer, self).__init__() | |
| 54 | |
| 55 self.src_dir = src_dir | |
| 56 self.build_dir = build_dir | |
| 57 self.req_tracker = req_tracker | |
| 58 | |
| 59 # Where still packed archives should be written to. If None, they are | |
| 60 # not saved, and are deleted immediately after unpacking. | |
| 61 self.download_dir = download_dir | |
| 62 | |
| 63 # Where still-packed .whl files should be written to. If None, they are | |
| 64 # written to the download_dir parameter. Separate to download_dir to | |
| 65 # permit only keeping wheel archives for pip wheel. | |
| 66 if wheel_download_dir: | |
| 67 wheel_download_dir = normalize_path(wheel_download_dir) | |
| 68 self.wheel_download_dir = wheel_download_dir | |
| 69 | |
| 70 # NOTE | |
| 71 # download_dir and wheel_download_dir overlap semantically and may | |
| 72 # be combined if we're willing to have non-wheel archives present in | |
| 73 # the wheelhouse output by 'pip wheel'. | |
| 74 | |
| 75 self.progress_bar = progress_bar | |
| 76 | |
| 77 # Is build isolation allowed? | |
| 78 self.build_isolation = build_isolation | |
| 79 | |
| 80 @property | |
| 81 def _download_should_save(self): | |
| 82 # type: () -> bool | |
| 83 # TODO: Modify to reduce indentation needed | |
| 84 if self.download_dir: | |
| 85 self.download_dir = expanduser(self.download_dir) | |
| 86 if os.path.exists(self.download_dir): | |
| 87 return True | |
| 88 else: | |
| 89 logger.critical('Could not find download directory') | |
| 90 raise InstallationError( | |
| 91 "Could not find or access download directory '%s'" | |
| 92 % display_path(self.download_dir)) | |
| 93 return False | |
| 94 | |
| 95 def prepare_linked_requirement( | |
| 96 self, | |
| 97 req, # type: InstallRequirement | |
| 98 session, # type: PipSession | |
| 99 finder, # type: PackageFinder | |
| 100 upgrade_allowed, # type: bool | |
| 101 require_hashes # type: bool | |
| 102 ): | |
| 103 # type: (...) -> AbstractDistribution | |
| 104 """Prepare a requirement that would be obtained from req.link | |
| 105 """ | |
| 106 # TODO: Breakup into smaller functions | |
| 107 if req.link and req.link.scheme == 'file': | |
| 108 path = url_to_path(req.link.url) | |
| 109 logger.info('Processing %s', display_path(path)) | |
| 110 else: | |
| 111 logger.info('Collecting %s', req) | |
| 112 | |
| 113 with indent_log(): | |
| 114 # @@ if filesystem packages are not marked | |
| 115 # editable in a req, a non deterministic error | |
| 116 # occurs when the script attempts to unpack the | |
| 117 # build directory | |
| 118 req.ensure_has_source_dir(self.build_dir) | |
| 119 # If a checkout exists, it's unwise to keep going. version | |
| 120 # inconsistencies are logged later, but do not fail the | |
| 121 # installation. | |
| 122 # FIXME: this won't upgrade when there's an existing | |
| 123 # package unpacked in `req.source_dir` | |
| 124 # package unpacked in `req.source_dir` | |
| 125 if os.path.exists(os.path.join(req.source_dir, 'setup.py')): | |
| 126 raise PreviousBuildDirError( | |
| 127 "pip can't proceed with requirements '%s' due to a" | |
| 128 " pre-existing build directory (%s). This is " | |
| 129 "likely due to a previous installation that failed" | |
| 130 ". pip is being responsible and not assuming it " | |
| 131 "can delete this. Please delete it and try again." | |
| 132 % (req, req.source_dir) | |
| 133 ) | |
| 134 req.populate_link(finder, upgrade_allowed, require_hashes) | |
| 135 | |
| 136 # We can't hit this spot and have populate_link return None. | |
| 137 # req.satisfied_by is None here (because we're | |
| 138 # guarded) and upgrade has no impact except when satisfied_by | |
| 139 # is not None. | |
| 140 # Then inside find_requirement existing_applicable -> False | |
| 141 # If no new versions are found, DistributionNotFound is raised, | |
| 142 # otherwise a result is guaranteed. | |
| 143 assert req.link | |
| 144 link = req.link | |
| 145 | |
| 146 # Now that we have the real link, we can tell what kind of | |
| 147 # requirements we have and raise some more informative errors | |
| 148 # than otherwise. (For example, we can raise VcsHashUnsupported | |
| 149 # for a VCS URL rather than HashMissing.) | |
| 150 if require_hashes: | |
| 151 # We could check these first 2 conditions inside | |
| 152 # unpack_url and save repetition of conditions, but then | |
| 153 # we would report less-useful error messages for | |
| 154 # unhashable requirements, complaining that there's no | |
| 155 # hash provided. | |
| 156 if is_vcs_url(link): | |
| 157 raise VcsHashUnsupported() | |
| 158 elif is_file_url(link) and is_dir_url(link): | |
| 159 raise DirectoryUrlHashUnsupported() | |
| 160 if not req.original_link and not req.is_pinned: | |
| 161 # Unpinned packages are asking for trouble when a new | |
| 162 # version is uploaded. This isn't a security check, but | |
| 163 # it saves users a surprising hash mismatch in the | |
| 164 # future. | |
| 165 # | |
| 166 # file:/// URLs aren't pinnable, so don't complain | |
| 167 # about them not being pinned. | |
| 168 raise HashUnpinned() | |
| 169 | |
| 170 hashes = req.hashes(trust_internet=not require_hashes) | |
| 171 if require_hashes and not hashes: | |
| 172 # Known-good hashes are missing for this requirement, so | |
| 173 # shim it with a facade object that will provoke hash | |
| 174 # computation and then raise a HashMissing exception | |
| 175 # showing the user what the hash should be. | |
| 176 hashes = MissingHashes() | |
| 177 | |
| 178 try: | |
| 179 download_dir = self.download_dir | |
| 180 # We always delete unpacked sdists after pip ran. | |
| 181 autodelete_unpacked = True | |
| 182 if req.link.is_wheel and self.wheel_download_dir: | |
| 183 # when doing 'pip wheel` we download wheels to a | |
| 184 # dedicated dir. | |
| 185 download_dir = self.wheel_download_dir | |
| 186 if req.link.is_wheel: | |
| 187 if download_dir: | |
| 188 # When downloading, we only unpack wheels to get | |
| 189 # metadata. | |
| 190 autodelete_unpacked = True | |
| 191 else: | |
| 192 # When installing a wheel, we use the unpacked | |
| 193 # wheel. | |
| 194 autodelete_unpacked = False | |
| 195 unpack_url( | |
| 196 req.link, req.source_dir, | |
| 197 download_dir, autodelete_unpacked, | |
| 198 session=session, hashes=hashes, | |
| 199 progress_bar=self.progress_bar | |
| 200 ) | |
| 201 except requests.HTTPError as exc: | |
| 202 logger.critical( | |
| 203 'Could not install requirement %s because of error %s', | |
| 204 req, | |
| 205 exc, | |
| 206 ) | |
| 207 raise InstallationError( | |
| 208 'Could not install requirement %s because of HTTP ' | |
| 209 'error %s for URL %s' % | |
| 210 (req, exc, req.link) | |
| 211 ) | |
| 212 abstract_dist = make_distribution_for_install_requirement(req) | |
| 213 with self.req_tracker.track(req): | |
| 214 abstract_dist.prepare_distribution_metadata( | |
| 215 finder, self.build_isolation, | |
| 216 ) | |
| 217 if self._download_should_save: | |
| 218 # Make a .zip of the source_dir we already created. | |
| 219 if not req.link.is_artifact: | |
| 220 req.archive(self.download_dir) | |
| 221 return abstract_dist | |
| 222 | |
| 223 def prepare_editable_requirement( | |
| 224 self, | |
| 225 req, # type: InstallRequirement | |
| 226 require_hashes, # type: bool | |
| 227 use_user_site, # type: bool | |
| 228 finder # type: PackageFinder | |
| 229 ): | |
| 230 # type: (...) -> AbstractDistribution | |
| 231 """Prepare an editable requirement | |
| 232 """ | |
| 233 assert req.editable, "cannot prepare a non-editable req as editable" | |
| 234 | |
| 235 logger.info('Obtaining %s', req) | |
| 236 | |
| 237 with indent_log(): | |
| 238 if require_hashes: | |
| 239 raise InstallationError( | |
| 240 'The editable requirement %s cannot be installed when ' | |
| 241 'requiring hashes, because there is no single file to ' | |
| 242 'hash.' % req | |
| 243 ) | |
| 244 req.ensure_has_source_dir(self.src_dir) | |
| 245 req.update_editable(not self._download_should_save) | |
| 246 | |
| 247 abstract_dist = make_distribution_for_install_requirement(req) | |
| 248 with self.req_tracker.track(req): | |
| 249 abstract_dist.prepare_distribution_metadata( | |
| 250 finder, self.build_isolation, | |
| 251 ) | |
| 252 | |
| 253 if self._download_should_save: | |
| 254 req.archive(self.download_dir) | |
| 255 req.check_if_exists(use_user_site) | |
| 256 | |
| 257 return abstract_dist | |
| 258 | |
| 259 def prepare_installed_requirement( | |
| 260 self, | |
| 261 req, # type: InstallRequirement | |
| 262 require_hashes, # type: bool | |
| 263 skip_reason # type: str | |
| 264 ): | |
| 265 # type: (...) -> AbstractDistribution | |
| 266 """Prepare an already-installed requirement | |
| 267 """ | |
| 268 assert req.satisfied_by, "req should have been satisfied but isn't" | |
| 269 assert skip_reason is not None, ( | |
| 270 "did not get skip reason skipped but req.satisfied_by " | |
| 271 "is set to %r" % (req.satisfied_by,) | |
| 272 ) | |
| 273 logger.info( | |
| 274 'Requirement %s: %s (%s)', | |
| 275 skip_reason, req, req.satisfied_by.version | |
| 276 ) | |
| 277 with indent_log(): | |
| 278 if require_hashes: | |
| 279 logger.debug( | |
| 280 'Since it is already installed, we are trusting this ' | |
| 281 'package without checking its hash. To ensure a ' | |
| 282 'completely repeatable environment, install into an ' | |
| 283 'empty virtualenv.' | |
| 284 ) | |
| 285 abstract_dist = InstalledDistribution(req) | |
| 286 | |
| 287 return abstract_dist |
