> At this point I guess I would recommend adding five or so
> g2.cores8.ram36.disk20 flavor VPSs to WMCS, with between one and three
> RTX A6000 GPUs each, plus a 1TB SSD each, which should cost under
> $60k. That should allow for very widely multilingual models somewhere
> between GPT-3.5 and 4 performance with current training rates.

Having part of the cluster for this makes sense, even as what it is used
for changes over time.

> These models can be quantized into int4 weights which run on cell
> phones:
> https://github.com/rupeshs/alpaca.cpp/tree/linux-android-build-support
> It seems inevitable that we will someday include such LLMs with
> Internet-in-a-Box, and, why not also the primary mobile apps

Eventually, yes. A good reason to renew attention to mobile as a canonical
wiki experience.
Wikimedia-l mailing list -- wikimedia-l@lists.wikimedia.org, guidelines at: 
https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and 
Public archives at 
To unsubscribe send an email to wikimedia-l-le...@lists.wikimedia.org

Reply via email to