Multiverse Computing pushes its compressed AI models into the mainstream
Spanish startup launches API portal for compressed AI models that run locally, cutting cloud costs and counterparty risk.
Multiverse Computing, a Spanish startup, is pushing its quantum-inspired compression technology into the mainstream with two key launches: the CompactifAI chat app and a new self-serve API portal. The company specializes in compressing large models from major labs like OpenAI, Meta, DeepSeek, and Mistral into smaller versions that can run locally on devices. Their chat app features 'Gilda,' a tiny model that operates offline, offering privacy by keeping data on-device. However, it automatically switches to cloud-based models via an API (dubbed 'Ash Nazg') if a device lacks sufficient RAM, which negates the privacy benefit. The app itself has seen limited consumer adoption, with fewer than 5,000 downloads last month.
The strategic focus is clearly on the enterprise. The newly launched API portal gives developers direct access to Multiverse's compressed models, bypassing marketplaces like AWS. CEO Enrique Lizaso highlights features like real-time usage monitoring, giving businesses the transparency and control needed for production deployment. The core value proposition is cost reduction and reduced dependency on external cloud infrastructure, a pressing concern as VC firm Lux Capital warns of financial instability in the AI supply chain. Multiverse's latest compressed model, HyperNova 60B 2602, is based on OpenAI's gpt-oss-120b and claims to deliver faster, cheaper performance than the original, signaling that the capability gap between large and small models is narrowing.
- Launches self-serve API portal for compressed AI models from OpenAI, Meta, DeepSeek, and Mistral, offering direct access without AWS Marketplace.
- Latest model HyperNova 60B 2602, built on OpenAI's gpt-oss-120b, claims faster responses at lower cost than the original.
- CompactifAI app features tiny 'Gilda' model for local, offline use but switches to cloud via 'Ash Nazg' API on low-RAM devices, with under 5,000 downloads last month.
Why It Matters
Offers enterprises a path to reduce soaring AI compute costs and dependency on unstable cloud providers by deploying efficient, compressed models.