UltRAG: a Universal Simple Scalable Recipe for Knowledge Graph RAG
New method tackles AI hallucinations by letting models query Wikidata-scale graphs with 116M entities.
A research team from the University of Cambridge and Meta AI has published UltRAG, a novel framework that fundamentally changes how large language models interact with structured knowledge. The system addresses the persistent problem of AI hallucinations by enabling LLMs to execute complex queries on knowledge graphs—structured databases of facts and relationships—without requiring any retraining of the base model or specialized query modules. This represents a significant departure from traditional retrieval-augmented generation (RAG) approaches, which are primarily designed for document-based data and struggle with the multi-step reasoning required for graph traversal.
UltRAG's key innovation is its modular architecture that separates the language understanding capabilities of an LLM from the graph query execution. The framework uses readily available, pre-trained neural modules to interpret a user's natural language question, formulate a formal graph query, execute it against the knowledge base, and then synthesize the results into a coherent answer. This approach allows the system to handle "multi-hop" questions that require connecting information across multiple nodes in the graph, a task that has historically been challenging for AI systems.
In experimental evaluations, UltRAG demonstrated superior performance compared to existing state-of-the-art knowledge graph RAG solutions across standard benchmarks. Perhaps most impressively, the framework proved capable of scaling to real-world knowledge graphs of immense size, successfully interfacing with the full Wikidata database containing 116 million entities and 1.6 billion relations. The researchers report that this scalability comes with computational costs that are comparable to or lower than existing methods, making the approach both powerful and practical for enterprise applications where factual accuracy is paramount.
- Eliminates need for retraining: Uses off-the-shelf LLMs and query executors without fine-tuning
- Handles massive scale: Successfully queries Wikidata with 116M entities and 1.6B relations
- Superior performance: Achieves state-of-the-art results on knowledge graph question answering benchmarks
Why It Matters
Enables enterprises to build factually accurate AI assistants that can reason across massive, structured databases without costly model retraining.