Nidus: Externalized Reasoning for AI-Assisted Engineering
A new 'governance runtime' enforces engineering rules on every commit, letting AI agents build provably correct software.
Researchers from cybiont GmbH have introduced Nidus, a novel framework that externalizes and enforces software engineering methodology for AI-assisted development. The system acts as a 'governance runtime,' compiling organizational standards into constraint libraries called guidebooks. These constraints are then enforced through decidable evaluation on every code mutation before it is persisted, ensuring that engineering invariants like traced requirements and evidenced deliveries are mechanically verified, not just learned behaviors.
In a groundbreaking demonstration, Nidus was used in a self-hosting deployment where three large language models—Claude, Gemini, and Codex—collaborated to build a 100,000-line software system. Crucially, every commit was verified against a current set of proof obligations, meaning the system being built also governed the rules of its own construction. This creates a 'recursive self-governance' where the constraint surface itself constrains mutations to itself, permanently eliminating classes of unengineered output and creating a formal development history where every state satisfies all active obligations.
The paper outlines four key contributions: recursive self-governance, stigmergic coordination that routes AI agents without central control, proximal spec reinforcement that uses the specification as a reward function at inference time (no weight updates needed), and governance theater prevention. The latter ensures compliance evidence cannot be fabricated within the modeled mutation path. This approach shifts the burden of assurance from being internalized by AI models through training to being enforced by an external, verifiable mechanism, addressing a core challenge in reliable AI-assisted engineering.
- Nidus is a governance runtime that enforces engineering methodology (like the V-model) as decidable constraints verified on every code commit.
- In a self-hosting demo, LLMs (Claude, Gemini, Codex) built a 100k-line system under proof obligations, with the system governing its own construction.
- It prevents 'governance theater' by making compliance evidence unfabricatable and uses 'proximal spec reinforcement' where UNSAT verdicts shape AI behavior at inference.
Why It Matters
This could enable large-scale, provably correct software development with AI, moving from probabilistic outputs to verifiably engineered artifacts.