b8245
The latest update patches a critical server issue and adds new Windows CUDA and Vulkan builds.
The open-source project Llama.cpp, maintained by the ggml-org team, has rolled out a new release tagged b8245. This update primarily addresses a specific server-side bug (issue #20232) related to the creation of checkpoints immediately after processing mtmd (mixture of experts) chunks. The fix prevents potential instability or data corruption during long-running inference sessions, making the server component more robust for developers deploying local language models.
Beyond the core fix, the release significantly expands the availability of pre-compiled binaries, lowering the barrier to entry. New builds for Windows now include versions for CUDA 12.4 and CUDA 13.1 DLLs, as well as a Vulkan backend, joining the existing CPU, CUDA, SYCL, and HIP options. This broadens hardware support for users with NVIDIA GPUs and those leveraging alternative graphics APIs. The release also maintains its comprehensive suite of builds for macOS (Apple Silicon and Intel), Linux (CPU, Vulkan, ROCm), iOS, and openEuler, cementing Llama.cpp's position as a cross-platform powerhouse for efficient LLM inference.
- Fixes server bug #20232 preventing erroneous checkpoint creation after mtmd chunks, improving reliability.
- Adds new Windows pre-built binaries for CUDA 12.4, CUDA 13.1, and Vulkan GPU backends.
- Maintains wide platform support across macOS, Linux, iOS, and openEuler with various acceleration options.
Why It Matters
Developers get a more stable server for local LLMs and easier access to GPU acceleration, lowering deployment friction.