CURE:Circuit-Aware Unlearning for LLM-based Recommendation
New circuit-aware approach selectively edits model components to remove user data without breaking recommendations.
A research team led by Ziheng Chen and Jiali Cheng has introduced CURE (Circuit-Aware Unlearning for LLM-based Recommendation), a novel framework addressing the critical privacy challenge in AI-powered recommendation systems. As LLMs like GPT-4 and Claude become central to personalized recommendations, they accumulate sensitive user data that must be removable under regulations like GDPR. Current unlearning methods treat models as black boxes, applying uniform updates that create destructive gradient conflicts between forgetting unwanted data and retaining useful knowledge.
CURE's breakthrough approach involves reverse-engineering LLMs to identify specific "circuits"—computational subgraphs causally responsible for recommendation behaviors. The framework analyzes how individual attention heads, feed-forward layers, and embedding components contribute to both remembering and forgetting objectives. Modules are then categorized into three groups: forget-specific (updated to remove target data), retain-specific (protected to preserve general knowledge), and task-shared (carefully balanced). This surgical precision allows CURE to achieve 40% better unlearning effectiveness than baseline methods while maintaining over 90% recommendation accuracy on MovieLens and Amazon review datasets.
The framework represents a significant step toward trustworthy AI deployment in sensitive domains like healthcare and finance recommendations. By making the unlearning process transparent and controllable, CURE addresses both technical and regulatory challenges. The team's experiments demonstrate that their approach scales to billion-parameter models while providing audit trails—showing exactly which circuits were modified for compliance verification. This work bridges the gap between powerful LLM capabilities and practical privacy requirements that have hindered real-world adoption.
- Identifies specific "circuits" within LLMs responsible for recommendation behaviors, enabling surgical data removal
- Categorizes model components into forget-specific, retain-specific, and shared groups to prevent gradient conflicts
- Maintains >90% recommendation accuracy while achieving 40% better unlearning than existing methods on real datasets
Why It Matters
Enables compliant deployment of LLM recommenders in regulated industries by providing transparent, auditable data removal mechanisms.