Admittedly, HTTP Basic Auth is better than nothing, but it's hard to get
excited about setting that up and either managing an extra set of
credentials, or hooking it up to a single-signon system, when the SSH-based
authentication is already set up to control access to the source
repositories.

On March 16, 2019 1:31:49 AM UTC, Tzu-ping Chung <uranu...@gmail.com> wrote:
>
>
> On 15 Mar 2019, at 21:47, Simon Ruggier <simo...@gmail.com> wrote:
>
> The packages have to be available online for dependency links to work,
> yes, but they're not public: one needs to authenticate with an SSH key to
> clone each repository.
>
>
> Both --find-links and --index-url (or --extra-index-url) provide the
> possibility to secure the download with authentication (HTTP Basic Auth).
>
>
>
> On March 13, 2019 2:09:03 PM UTC, Tzu-ping Chung <uranu...@gmail.com>
> wrote:
>>
>>
>> On 6 Mar 2019, at 03:53, Simon <simo...@gmail.com> wrote:
>>>
>>> I hope it's not an issue that I'm replying to a month-old thread. I 
>>> reviewed the previous discussion to try to avoid duplicating any of it.
>>>
>>> When using pip with PyPI, calling pip a second time is much quicker than 
>>> the first time, because it verifies that the requirements, including 
>>> version constraints, are satisfied in the target environment and doesn't 
>>> needlessly reinstall stuff.
>>>
>>> Dependency links allowed the same behaviour to be implemented for private 
>>> packages with dependencies on other private repositories: given a 
>>> requirement B >= 3 and a dependency link that B was available from, pip 
>>> could check if the environment already includes a package B with a new 
>>> enough version, and only use the dependency link as a fallback if the 
>>> requirement isn't already satisfied.
>>>
>>> URL specifiers aren't useful for providing a fallback location to get a 
>>> package from, because using one prevents the package from specifying a 
>>> version constraint in the same way that was possible with dependency links, 
>>> or with normal requirements available from PyPI. Curiously, discussion of 
>>> version constraints in this thread has focused on how nonsensical it would 
>>> be to compare them to the specifying URL, ignoring the possibility of 
>>> comparing the constraint with the target environment.
>>>
>>> The loss of this functionality means that anyone who was previously using 
>>> pip to automatically install private packages with private dependencies now 
>>> has to either forgo automatic dependency management (a large part of why 
>>> one would use a package manager to begin with) in favour of recursively 
>>> specified requirements files, publish their private packages somewhere so 
>>> that pip can find them, or stick with pip 18.1 for now.
>>>
>>
>> Wouldn’t you still need to “publish the private packages somewhere” for 
>> dependency links to work? `setup.py sdist` with `pip --find-links` can get 
>> you very far; the only differences IMO is you have to provide a proper 
>> package, and write a simple HTML file to point to it.
>>
>>
>> --
>>> Distutils-SIG mailing list -- distutils-sig@python.org
>>> To unsubscribe send an email to distutils-sig-le...@python.org
>>> https://mail.python.org/mailman3/lists/distutils-sig.python.org/
>>> Message archived at 
>>> https://mail.python.org/archives/list/distutils-sig@python.org/message/BALD2PVKGHBBWIKNYTZGGF6LHEXI7O26/
>>>
>>
--
Distutils-SIG mailing list -- distutils-sig@python.org
To unsubscribe send an email to distutils-sig-le...@python.org
https://mail.python.org/mailman3/lists/distutils-sig.python.org/
Message archived at 
https://mail.python.org/archives/list/distutils-sig@python.org/message/UVXDLI6ICWU4HGYTGVUWIAGGZH4LP2ZZ/

Reply via email to