Add llama.cpp as a CPU-optimized ServingRuntime for LLM inference
April 2, 2026 ยท #5334
Go
Difficulty: Medium
Parent Repository
kserve/kserve
Go repository
5,285 1,423
kserve/kserve
Go repository
Sign in required
Authenticate to use favourites & bookmarks