Hi,

an update of llama.cpp is in the works, it was just stalled by two bugs
in the ggml build.

It should happen by the weekend, latest.

Best,
Christian

On 2026-03-26 18:46, Philipp Klaus Krause wrote:
> Package: llama.cpp
> Version: 8064+dfsg-2
> Severity: normal
> X-Debbugs-Cc: [email protected]
> 
> Dear Maintainer,
> 
> I am trying to use a Qwen3-coder model with llama.cpp and opencode. However, I
> see frequent crashes of llama-server. One example can be found below. I 
> suspect
> that this is upstream llama.cpp bug #19304 (https://github.com/ggml-
> org/llama.cpp/issues/19304), which was fixed about a week after the release of
> the currently packaged version, though the bug reports there used different
> Qwen-Coder variants.
> 
> […]
> slot update_slots: id  3 | task 32108 | prompt done, n_tokens = 38850,
> batch.n_tokens = 17
> slot init_sampler: id  3 | task 32108 | init sampler, took 5.85 ms, tokens:
> text = 38850, total = 38850
> slot update_slots: id  3 | task 32108 | created context checkpoint 3 of 8
> (pos_min = 38832, pos_max = 38832, size = 75.376 MiB)
> [New LWP 3895376]
> [New LWP 3895375]
> [New LWP 3895374]
> [New LWP 3895373]
> [New LWP 3895372]
> [New LWP 3895371]
> [New LWP 3895370]
> [New LWP 3895369]
> [New LWP 3895368]
> [New LWP 3895367]
> [New LWP 3895366]
> [New LWP 3895365]
> [New LWP 3895364]
> [New LWP 3895363]
> [New LWP 3895362]
> [New LWP 3895128]
> [New LWP 3895127]
> [New LWP 3895126]
> [New LWP 3895125]
> [New LWP 3895124]
> [New LWP 3895123]
> [New LWP 3895122]
> [New LWP 3895121]
> [New LWP 3895120]
> [New LWP 3895119]
> [New LWP 3895118]
> [New LWP 3895117]
> [New LWP 3895116]
> [New LWP 3895115]
> [New LWP 3895114]
> [New LWP 3895113]
> [New LWP 3895112]
> [New LWP 3895111]
> [New LWP 3895110]
> [New LWP 3895109]
> [New LWP 3895108]
> [New LWP 3895107]
> [New LWP 3895106]
> [New LWP 3895105]
> [New LWP 3895104]
> [New LWP 3895103]
> [New LWP 3895102]
> [New LWP 3895101]
> [New LWP 3895100]
> [New LWP 3895099]
> [New LWP 3895098]
> [New LWP 3895097]
> [New LWP 3895096]
> [New LWP 3895094]
> [New LWP 3895093]
> 
> This GDB supports auto-downloading debuginfo from the following URLs:
>   <https://debuginfod.debian.net>
> Enable debuginfod for this session? (y or [n]) [answered N; input not from
> terminal]
> Debuginfod has been disabled.
> To make this setting permanent, add 'set debuginfod enabled off' to .gdbinit.
> [Thread debugging using libthread_db enabled]
> Using host libthread_db library "/usr/lib/x86_64-linux-gnu/libthread_db.so.1".
> __syscall_cancel_arch () at
> ../sysdeps/unix/sysv/linux/x86_64/syscall_cancel.S:56
> ⚠️ warning: 56  ../sysdeps/unix/sysv/linux/x86_64/syscall_cancel.S: Datei oder
> Verzeichnis nicht gefunden
> #0  __syscall_cancel_arch () at
> ../sysdeps/unix/sysv/linux/x86_64/syscall_cancel.S:56
> 56      in ../sysdeps/unix/sysv/linux/x86_64/syscall_cancel.S
> #1  0x00007f9b9dc9be64 in __internal_syscall_cancel (a1=<optimized out>,
> a2=<optimized out>, a3=<optimized out>, a4=<optimized out>, a5=a5@entry=0,
> a6=a6@entry=0, nr=61) at ./nptl/cancellation.c:49
> ⚠️ warning: 49  ./nptl/cancellation.c: Datei oder Verzeichnis nicht gefunden
> #2  0x00007f9b9dc9bead in __syscall_cancel (a1=<optimized out>, a2=<optimized
> out>, a3=<optimized out>, a4=<optimized out>, a5=a5@entry=0, a6=a6@entry=0,
> nr=61) at ./nptl/cancellation.c:75
> 75      in ./nptl/cancellation.c
> #3  0x00007f9b9dd07c07 in __GI___wait4 (pid=<optimized out>,
> stat_loc=<optimized out>, options=<optimized out>, usage=<optimized out>) at
> ../sysdeps/unix/sysv/linux/wait4.c:30
> ⚠️ warning: 30  ../sysdeps/unix/sysv/linux/wait4.c: Datei oder Verzeichnis
> nicht gefunden
> #4  0x00007f9b9ec6aa53 in ggml_print_backtrace () from /usr/lib/x86_64-linux-
> gnu/libggml-base.so.0
> #5  0x00007f9b9ec79c3f in ?? () from /usr/lib/x86_64-linux-gnu/libggml-
> base.so.0
> #6  0x00007f9b9debb5fa in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6
> #7  0x00007f9b9dea7749 in std::terminate() () from /usr/lib/x86_64-linux-
> gnu/libstdc++.so.6
> #8  0x00007f9b9debb898 in __cxa_throw () from /usr/lib/x86_64-linux-
> gnu/libstdc++.so.6
> #9  0x00007f9b9e2796ea in ?? () from /usr/lib/x86_64-linux-
> gnu/llama/libllama.so.0
> #10 0x00007f9b9e2cd05c in llama_grammar_accept_impl(llama_grammar&, int) ()
> from /usr/lib/x86_64-linux-gnu/llama/libllama.so.0
> #11 0x000055c375409ba8 in ?? ()
> #12 0x000055c375277c94 in ?? ()
> #13 0x000055c3752c125e in ?? ()
> #14 0x000055c3751d5209 in ?? ()
> #15 0x00007f9b9dc33f75 in __libc_start_call_main
> (main=main@entry=0x55c3751d1390, argc=argc@entry=19,
> argv=argv@entry=0x7ffc39c3f588) at ../sysdeps/nptl/libc_start_call_main.h:58
> ⚠️ warning: 58  ../sysdeps/nptl/libc_start_call_main.h: Datei oder Verzeichnis
> nicht gefunden
> #16 0x00007f9b9dc34027 in __libc_start_main_impl (main=0x55c3751d1390, 
> argc=19,
> argv=0x7ffc39c3f588, init=<optimized out>, fini=<optimized out>,
> rtld_fini=<optimized out>, stack_end=0x7ffc39c3f578) at 
> ../csu/libc-start.c:360
> ⚠️ warning: 360 ../csu/libc-start.c: Datei oder Verzeichnis nicht gefunden
> #17 0x000055c3751da7f1 in ?? ()
> [Inferior 1 (process 3895092) detached]
> terminate called after throwing an instance of 'std::runtime_error'
>   what():  Unexpected empty grammar stack after accepting piece: =list (40972)
> ./server-example: Zeile 11: 3895092 Abgebrochen                llama-server
> --model models/Qwen3-Coder-Next-Q4_K_S.gguf --ctx-size 131072 --alias
> "Qwen3-Coder-Next" --seed 3407 --temp 1.0 --top-p 0.95 --min-p 0.01 --top-k 40
> --port 8081
> 
> 
> -- System Information:
> Debian Release: forky/sid
>   APT prefers unstable
>   APT policy: (500, 'unstable'), (500, 'testing')
> Architecture: amd64 (x86_64)
> Foreign Architectures: i386
> 
> Kernel: Linux 6.19.8+deb14-amd64 (SMP w/32 CPU threads; PREEMPT)
> Locale: LANG=de_DE.UTF-8, LC_CTYPE=de_DE.UTF-8 (charmap=UTF-8), LANGUAGE not 
> set
> Shell: /bin/sh linked to /usr/bin/dash
> Init: systemd (via /run/systemd/system)
> LSM: AppArmor: enabled
> 
> Versions of packages llama.cpp depends on:
> ii  llama.cpp-tools  8064+dfsg-2
> 
> Versions of packages llama.cpp recommends:
> ii  llama.cpp-tools-extra  8064+dfsg-2
> ii  python3-gguf           8064+dfsg-2
> 
> Versions of packages llama.cpp suggests:
> pn  llama.cpp-examples  <none>
> 
> -- no debconf information

Reply via email to