Eval bug: llama-server reproducibly enters degenerate per-slot generation state after contaminated prompt; returns 1-4 token completions until context reset
May 8, 2026 ยท #22828
cpp
Difficulty: Medium
Labels
bug-unconfirmed
Parent Repository
ggml-org/llama.cpp
cpp repository
109,099 17,966