b8356
Latest commit patches a division-by-zero vulnerability in the cutting-edge IQ4_NL quantization format.
The open-source powerhouse behind llama.cpp, ggml-org, has rolled out a new update with commit b8356. This release primarily addresses a critical bug in the code for the IQ4_NL quantization format, implementing a guard to prevent a division-by-zero error when the variable `sumq2` equals zero. This fix enhances the stability of models quantized with this advanced, memory-efficient format, which is key for running large language models on consumer hardware.
Alongside the bug fix, the team has published a comprehensive suite of pre-built binaries, significantly lowering the barrier to entry for developers. The release supports an impressive range of operating systems and hardware accelerators, from macOS on Apple Silicon and Intel to Windows with CUDA 12.4/13.1, Vulkan, and SYCL backends. It also includes builds for Linux with Vulkan and ROCm 7.2 support, and specialized versions for Huawei's openEuler OS and Ascend AI processors (310p, 910b). This broad compatibility underscores llama.cpp's role as a universal tool for efficient, local AI inference.
The commit (b9da444) was automatically released by GitHub Actions and is cryptographically signed, ensuring its authenticity. By providing these ready-to-use binaries, the llama.cpp project continues to democratize access to high-performance LLM inference, allowing developers and researchers to bypass complex build processes and deploy models directly on everything from laptops to specialized servers.
- Critical bug fix for IQ4_NL quantization prevents a division-by-zero error (guard against sumq2 being 0).
- Massive cross-platform support with 24 pre-built assets for macOS, Windows, Linux, and openEuler on x64, arm64, and s390x architectures.
- Includes builds for major GPU compute platforms: CUDA 12.4/13.1, Vulkan, ROCm 7.2, SYCL, HIP, and OpenVINO.
Why It Matters
Ensures stability for cutting-edge model compression and simplifies deployment of efficient local LLMs across virtually any hardware stack.