Package: wnpp Severity: wishlist Owner: Mo Zhou <lu...@debian.org> X-Debbugs-Cc: debian-de...@lists.debian.org, debian...@lists.debian.org
* Package name : transformers Upstream Contact: HuggingFace * URL : https://github.com/huggingface/transformers * License : Apache-2.0 Description : State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow I've been using this for a while. This package provides a convenient way for people to download and run an LLM locally. Basically, if you want to run an instruct fine-tuned large language model with 7B parameters, you will need at least 16GB of CUDA memory for inference in half/bfloat16 precision. I have not tried to run any LLM with > 3B parameters with CPU ... that can be slow. LLaMa.cpp is a good choice for running LLM on CPU, but that library supports less models than this one. Meanwhile, the cpp library only supports inference. I don't know how many dependencies are still missing, but that should not be too much. Jax and TensorFlow are optional dependencies so they can be missing from our archive. But anyway, I think running a large language model locally with Debian packages will be interesting. The CUDA version of PyTorch is already in the NEW queue. That said, this is actually a very comprehensive library, which provides far more functionalities than running LLMs. Thank you for using reportbug