Recently one of the Libcloud contributors (Eis-D-Z) published various improvements to our price scraping scripts and added some new ones - https://github.com/apache/libcloud/pulls/Eis-D-Z.
I think it would now make sense to run those scraping scripts on a continuous basis as part of our CI (e.g. once a day) and publish the generated file to some well known location (e.g. public read-only S3 bucket). In fact, that was also the plan when we originally added libcloud.pricing.download_pricing_file function and related functionality quite a long time ago. IIRC, the plan was to include an auto-generated pricing file directly inside the git repo, but this is more complicated and I would need to contact the ASF infra team if they even allow something like that (updating and committing a change as a bot user on our CI - Travis CI). So for now, I will probably just publish this auto-generated pricing.json file to a public read-only S3 bucket (I will make sure to set up correct rate limits and alerts to prevent abuse, even though the pricing file itself is quite small). What do other people think?