Those files have now been made public.

I will publish a blog post with some details on that in the near future.

On Sat, Jul 11, 2020 at 8:48 PM Tomaz Muraus <to...@apache.org> wrote:

> I added some information on this new behavior here -
> https://github.com/apache/libcloud/blob/f122600d2adf181a9b100cdd552cd02979c5b1b9/docs/compute/pricing.rst#downloading-latest-pricing-data-from-an-s3-bucket
>
> Keep in mind that those 3 files are not public yet. I plan to make them
> public and read-only in the near future once the rate limits are sorted out.
>
> On Sat, Jul 11, 2020 at 4:07 PM Tomaz Muraus <to...@apache.org> wrote:
>
>> Yeah, I would actually prefer a git repository so everything is version
>> controlled, etc., but I went with the fastest and simplest approach
>> possible.
>>
>> I'm not exactly sure what the ASF rules are for something like that (I
>> would need to ask ASF infra team to create a new repo, create a bot account
>> which we could use in our CI, etc.) and that would likely take much longer
>> than the approach I went with.
>>
>> As far as libraries such as pytz (and to some extent also certifi) go - I
>> would say it's a slightly different there - time zones tend to change much
>> less frequently than provider pricings so publishing a new library package
>> every now and then is probably sufficient.
>>
>> On Sat, Jul 11, 2020 at 2:38 PM Samuel Marks <samuelma...@gmail.com>
>> wrote:
>>
>>> The other solution is to create a new git repository just for frequently
>>> updated files like this one… I mean we don't want to end up like pytz do
>>> we?
>>>
>>> PS: A good thing about pytz is other languages literally just parse
>>> pytz's
>>> list for their own timezone implementation. No Python. Easy! - With this
>>> being in JSON, I could imagine using Terraform libraries in Go; instead
>>> of
>>> Libcloud; to do multicloud, and use this costing system to say where and
>>> when.
>>>
>>> Samuel Marks
>>> Charity <https://sydneyscientific.org> | consultancy <
>>> https://offscale.io>
>>> | open-source <https://github.com/offscale> | LinkedIn
>>> <https://linkedin.com/in/samuelmarks>
>>>
>>>
>>> On Thu, Jul 2, 2020 at 9:51 PM Jay Rolette <role...@infinite.io> wrote:
>>>
>>> > Same here!
>>> >
>>> > Thanks,
>>> > Jay
>>> >
>>> > On Wed, Jul 1, 2020 at 12:45 PM Francisco Ros <fj...@doalitic.com>
>>> wrote:
>>> >
>>> > > Hey Tomaz,
>>> > >
>>> > > I'd really love to see this :-)
>>> > >
>>> > > Thanks,
>>> > > Francisco
>>> > >
>>> > > > El 1 jul 2020, a las 12:00, Tomaz Muraus <to...@apache.org>
>>> escribió:
>>> > > >
>>> > > > Recently one of the Libcloud contributors (Eis-D-Z) published
>>> various
>>> > > > improvements to our price scraping scripts and added some new ones
>>> -
>>> > > > https://github.com/apache/libcloud/pulls/Eis-D-Z.
>>> > > >
>>> > > > I think it would now make sense to run those scraping scripts on a
>>> > > > continuous basis as part of our CI (e.g. once a day) and publish
>>> the
>>> > > > generated file to some well known location (e.g. public read-only
>>> S3
>>> > > > bucket).
>>> > > >
>>> > > > In fact, that was also the plan when we originally
>>> > > > added libcloud.pricing.download_pricing_file function and related
>>> > > > functionality quite a long time ago.
>>> > > >
>>> > > > IIRC, the plan was to include an auto-generated pricing file
>>> directly
>>> > > > inside the git repo, but this is more complicated and I would need
>>> to
>>> > > > contact the ASF infra team if they even allow something like that
>>> > > (updating
>>> > > > and committing a change as a bot user on our CI - Travis CI).
>>> > > >
>>> > > > So for now, I will probably just publish this auto-generated
>>> > pricing.json
>>> > > > file to a public read-only S3 bucket (I will make sure to set up
>>> > correct
>>> > > > rate limits and alerts to prevent abuse, even though the pricing
>>> file
>>> > > > itself is quite small).
>>> > > >
>>> > > > What do other people think?
>>> > >
>>> > >
>>> >
>>>
>>

Reply via email to