b8196
Critical fix enables support for models with vocabularies exceeding 100,000 tokens, including Google's new Gemma 3.
The open-source project llama.cpp, maintained by ggml-org, has released a critical technical update (commit b8196) that fixes a previously overlooked bug in tensor shape formatting. The issue was in the `llama_format_tensor_shape` function, which only allocated 5 digits for tensor dimensions, causing failures when loading modern LLMs with massive vocabulary sizes. This directly impacts support for newly released models like Google's Gemma 3, which has a vocabulary of 262,208 tokens—a number requiring 6 digits. The fix is essential for the ecosystem, as llama.cpp is a foundational C++ library for efficient inference of models like Llama, Mistral, and Gemma on consumer hardware.
The commit, automatically released via GitHub Actions, ensures broad compatibility across llama.cpp's extensive platform matrix. This includes builds for Apple Silicon and Intel macOS, iOS, various Linux distributions (with CPU, Vulkan, and ROCm 7.2 support), and multiple Windows configurations (CPU, CUDA 12/13, Vulkan, SYCL, and HIP). The patch, while small, is a necessary infrastructure update that maintains the library's role as a versatile, high-performance backend for the open-source AI community, enabling developers and researchers to run the latest models locally without interruption.
- Fixes a tensor dimension bug in `llama_format_tensor_shape` that blocked models with vocabularies > 99,999 tokens.
- Enables immediate support for Google's newly announced Gemma 3 model, which has a 262,208-token vocabulary.
- Update is rolled out across all 23+ platform-specific builds, including CUDA, Vulkan, ROCm, and CPU backends.
Why It Matters
This maintenance fix is crucial for developers relying on llama.cpp to run the latest open-source LLMs locally on their own hardware.