Recently one of the Libcloud contributors (Eis-D-Z) published various
improvements to our price scraping scripts and added some new ones -
https://github.com/apache/libcloud/pulls/Eis-D-Z.

I think it would now make sense to run those scraping scripts on a
continuous basis as part of our CI (e.g. once a day) and publish the
generated file to some well known location (e.g. public read-only S3
bucket).

In fact, that was also the plan when we originally
added libcloud.pricing.download_pricing_file function and related
functionality quite a long time ago.

IIRC, the plan was to include an auto-generated pricing file directly
inside the git repo, but this is more complicated and I would need to
contact the ASF infra team if they even allow something like that (updating
and committing a change as a bot user on our CI - Travis CI).

So for now, I will probably just publish this auto-generated pricing.json
file to a public read-only S3 bucket (I will make sure to set up correct
rate limits and alerts to prevent abuse, even though the pricing file
itself is quite small).

What do other people think?

Reply via email to