b8828
The popular open-source project now supports Google's latest Gemma 4 model across 27 platform builds.
The open-source community behind llama.cpp has released a significant update with commit b8828, adding comprehensive support for Google's recently announced Gemma 4 model. This integration marks a major expansion of the popular local AI inference framework's capabilities, allowing developers to run Google's latest 9B-parameter model alongside existing Meta Llama models. The update includes specific model type detection for Gemma 4, ensuring proper handling of its architecture and optimization for local deployment across diverse hardware configurations.
The release provides 27 different pre-built binaries covering virtually every major platform combination. For macOS users, there are builds for both Apple Silicon (arm64) and Intel (x64) architectures, including a special KleidiAI-enabled version for enhanced performance. Windows users get options ranging from standard CPU builds to specialized versions with CUDA 12.4, CUDA 13.1, Vulkan, SYCL, and HIP support. Linux distributions include Ubuntu builds with CPU, Vulkan, ROCm 7.2, and OpenVINO backends, while openEuler users get specialized builds for Huawei's Ascend 310p and 910b hardware with ACL Graph acceleration.
This update represents a significant milestone for the local AI ecosystem, as llama.cpp has become the de facto standard for running large language models on consumer hardware. The addition of Gemma 4 support gives developers another powerful option in their toolkit, particularly valuable for applications requiring Google's specific architectural strengths or for those working in environments where Google's models are preferred. The comprehensive platform coverage ensures that whether developers are working on iOS mobile applications, Windows gaming PCs, or Linux servers, they have optimized builds available for their specific hardware configuration.
- Adds official Gemma 4 model support to llama.cpp with proper type detection
- Provides 27 different platform builds including macOS Apple Silicon, Windows CUDA, and Linux Vulkan
- Includes specialized openEuler builds for Huawei Ascend hardware with ACL Graph acceleration
Why It Matters
Developers gain another powerful local AI option with Google's architecture, expanding what's possible on consumer hardware without cloud dependencies.