So, everyone wants to host their own (copy of a) large language model
(LLM), but many academic institutions can't spin up multiple LLMs
simultaneously, in perpetuity, nor do I believe the Scientific Funding
Agencies in each country would want to pay for everyone to get a GPU
cluster just to host 500+ copies of tomorrow's version of LLAMA-2(ish).

Are you aware of any effort proposing or planning to host LLMs for use by
researchers in some shared infrastructure? After all, hosting the LLM costs
the same per hour whether 1, 3 or 20 people are calling it, and at most
academic institutions usage would be a little bursty.

Best,
Amanda Stent

--
(they/she)
Director, Davis Institute for AI
Professor, Computer Science
Colby College

Follow the Davis Institute for AI here
<https://web.colby.edu/davisinstitute/>
Want to meet? Calendly - Amanda Stent
<https://calendly.com/amandastentcolby>
_______________________________________________
Corpora mailing list -- corpora@list.elra.info
https://list.elra.info/mailman3/postorius/lists/corpora.list.elra.info/
To unsubscribe send an email to corpora-le...@list.elra.info

Reply via email to