b9072
SPIR-V shadowing fix lands in the latest llama.cpp release
The latest tag b9072 for ggml-org/llama.cpp addresses a Vulkan SPIR-V shadowing bug (issue #22760) that could cause rendering artifacts or crashes in local LLM inference when using the Vulkan backend. This fix, verified with GPG signature, is crucial for users leveraging Vulkan for cross-platform GPU acceleration, as SPIR-V shadowing errors can lead to incorrect shader behavior. The release ensures smoother operation on systems without NVIDIA CUDA or AMD ROCm support, particularly on Linux and Windows with integrated or non-native GPUs.
Beyond the bug fix, b9072 significantly expands platform support. It provides pre-built binaries for macOS (Apple Silicon with optional KleidiAI, Intel x64), Linux (x64, arm64, even s390x mainframe), Windows (x64 and arm64 CPU, CUDA 12 & 13, Vulkan, SYCL, HIP), Android (arm64 CPU), and openEuler (custom x86 and aarch64 builds with ACL Graph). This breadth makes llama.cpp accessible on everything from smartphones to enterprise servers. For developers and researchers running models like LLaMA, Mistral, or Gemma locally, this release reduces setup friction and increases hardware compatibility, reinforcing llama.cpp's position as the go-to solution for on-device AI.
- Fixes SPIR-V shadowing bug (#22760) in Vulkan backend for correct shader execution
- Supports 20+ platform/backend combos: CPU, CUDA 12 & 13, ROCm 7.2, Vulkan, SYCL, OpenVINO, HIP, and ACL Graph
- Available on macOS, Windows, Linux, Android, and openEuler across x64, arm64, and s390x architectures
Why It Matters
Stabilizes Vulkan-based local LLM inference, broadening hardware support for on-device AI deployment.