Adaptive Memory Admission Control for LLM Agents
New framework uses five interpretable factors to decide what AI agents should remember, slashing latency by nearly a third.
A team of researchers has introduced Adaptive Memory Admission Control (A-MAC), a novel framework designed to solve a critical bottleneck in AI agent development: uncontrolled memory growth. Current LLM-based agents, which rely on long-term memory for multi-session tasks, either hoard vast amounts of unreliable conversational data or use opaque, expensive LLM calls to manage memory. A-MAC reframes memory admission as a structured decision problem, combining lightweight rule-based feature extraction with a single, optimized LLM call to assess an item's future utility. This hybrid approach provides the transparency and auditability that purely LLM-driven policies lack.
Technically, A-MAC decomposes the value of a potential memory into five complementary and interpretable factors: future utility, factual confidence, semantic novelty, temporal recency, and a content type prior. The system learns domain-adaptive admission policies through cross-validated optimization of these factors. In experiments on the LoCoMo benchmark, this method proved highly effective, achieving a superior precision-recall tradeoff with an F1 score of 0.583. Crucially, it also reduced system latency by 31% compared to state-of-the-art LLM-native memory systems. An ablation study identified the content type prior—a statistical prior over different kinds of information—as the most influential factor for reliable admission. The findings establish explicit, interpretable admission control as a key design principle for building scalable and trustworthy AI agents.
- A-MAC framework improves memory F1 score to 0.583 on the LoCoMo benchmark while reducing system latency by 31%.
- It evaluates memories using five interpretable factors: future utility, factual confidence, semantic novelty, temporal recency, and content type prior.
- The hybrid design combines rule-based features with one optimized LLM call, moving beyond opaque and costly fully LLM-driven memory policies.
Why It Matters
Enables more efficient, scalable, and auditable AI agents by giving developers precise control over what their systems remember and forget.