> > At this point I guess I would recommend adding five or so > g2.cores8.ram36.disk20 flavor VPSs to WMCS, with between one and three > RTX A6000 GPUs each, plus a 1TB SSD each, which should cost under > $60k. That should allow for very widely multilingual models somewhere > between GPT-3.5 and 4 performance with current training rates. >
Having part of the cluster for this makes sense, even as what it is used for changes over time. > These models can be quantized into int4 weights which run on cell > phones: > https://github.com/rupeshs/alpaca.cpp/tree/linux-android-build-support > It seems inevitable that we will someday include such LLMs with > Internet-in-a-Box, and, why not also the primary mobile apps > Eventually, yes. A good reason to renew attention to mobile as a canonical wiki experience.
_______________________________________________ Wikimedia-l mailing list -- email@example.com, guidelines at: https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and https://meta.wikimedia.org/wiki/Wikimedia-l Public archives at https://firstname.lastname@example.org/message/R2KSXD3VE4MPGSIIYCPUMFXFPIY7D5CH/ To unsubscribe send an email to wikimedia-l-le...@lists.wikimedia.org