Bug#1063673: ITP: llama.cpp -- Inference of Meta's LLaMA model (and others) in pure C/C++

2024-03-10 Thread Petter Reinholdtsen
[Christian Kastner] > I'm open for better ideas, though. I find in general that programs written with run time selection of optimizations are far superiour to per host compilations, at least from a system administration viewpoint. I guess such approach would require rewriting llama.cpp, and have

Bug#1063673: ITP: llama.cpp -- Inference of Meta's LLaMA model (and others) in pure C/C++

2024-03-09 Thread Christian Kastner
Hey Ptter, On 2024-03-08 20:21, Petter Reinholdtsen wrote: > [Christian Kastner 2024-02-13] >> I'll push a first draft soon, though it will definitely not be >> upload-ready for the above reasons. > > Where can I find the first draft? I've discarded the simple package and now plan another

Bug#1063673: ITP: llama.cpp -- Inference of Meta's LLaMA model (and others) in pure C/C++

2024-03-08 Thread Petter Reinholdtsen
[Christian Kastner 2024-02-13] > I'll push a first draft soon, though it will definitely not be > upload-ready for the above reasons. Where can I find the first draft? -- Happy hacking Petter Reinholdtsen

Bug#1063673: ITP: llama.cpp -- Inference of Meta's LLaMA model (and others) in pure C/C++

2024-02-13 Thread Christian Kastner
Hi Petter, On 2024-02-13 08:36, Petter Reinholdtsen wrote: > I tried building the CPU edition on one machine and run it on another, > and experienced illegal instruction exceptions. I suspect this mean one > need to be careful when selecting build profile to ensure it work on all > supported

Bug#1063673: ITP: llama.cpp -- Inference of Meta's LLaMA model (and others) in pure C/C++

2024-02-12 Thread Petter Reinholdtsen
I tried building the CPU edition on one machine and run it on another, and experienced illegal instruction exceptions. I suspect this mean one need to be careful when selecting build profile to ensure it work on all supported Debian platforms. I would be happy to help getting this up and

Bug#1063673: ITP: llama.cpp -- Inference of Meta's LLaMA model (and others) in pure C/C++

2024-02-10 Thread Christian Kastner
Package: wnpp Severity: wishlist Owner: Christian Kastner X-Debbugs-Cc: debian-de...@lists.debian.org, debian...@lists.debian.org * Package name: llama.cpp Version : b2116 Upstream Author : Georgi Gerganov * URL : https://github.com/ggerganov/llama.cpp * License