Paper List
-
Developing the PsyCogMetrics™ AI Lab to Evaluate Large Language Models and Advance Cognitive Science
This paper addresses the critical gap between sophisticated LLM evaluation needs and the lack of accessible, scientifically rigorous platforms that in...
-
Equivalence of approximation by networks of single- and multi-spike neurons
This paper resolves the fundamental question of whether single-spike spiking neural networks (SNNs) are inherently less expressive than multi-spike SN...
-
The neuroscience of transformers
提出了Transformer架构与皮层柱微环路之间的新颖计算映射,连接了现代AI与神经科学。
-
Framing local structural identifiability and observability in terms of parameter-state symmetries
This paper addresses the core challenge of systematically determining which parameters and states in a mechanistic ODE model can be uniquely inferred ...
-
Leveraging Phytolith Research using Artificial Intelligence
This paper addresses the critical bottleneck in phytolith research by automating the labor-intensive manual microscopy process through a multimodal AI...
-
Neural network-based encoding in free-viewing fMRI with gaze-aware models
This paper addresses the core challenge of building computationally efficient and ecologically valid brain encoding models for naturalistic vision by ...
-
Scalable DNA Ternary Full Adder Enabled by a Competitive Blocking Circuit
This paper addresses the core bottleneck of carry information attenuation and limited computational scale in DNA binary adders by introducing a scalab...
-
ELISA: An Interpretable Hybrid Generative AI Agent for Expression-Grounded Discovery in Single-Cell Genomics
This paper addresses the critical bottleneck of translating high-dimensional single-cell transcriptomic data into interpretable biological hypotheses ...
D-MEM: Dopamine-Gated Agentic Memory via Reward Prediction Error Routing
UC San Diego | Carnegie Mellon University
30秒速读
IN SHORT: This paper addresses the fundamental scalability bottleneck in LLM agentic memory systems: the O(N²) computational complexity and unbounded API token costs caused by processing every user utterance through expensive memory evolution pipelines, regardless of information value.
核心创新
- Methodology Introduces D-MEM, a bio-inspired architecture implementing dopamine-gated fast/slow routing based on Agentic Reward Prediction Error (RPE), decoupling short-term interaction from long-term cognitive restructuring.
- Methodology Develops the LoCoMo-Noise benchmark with systematic 75% noise injection (Filler: 40%, Status: 30%, Tangent: 30%) to simulate real-world conversational dynamics and evaluate memory robustness.
- Methodology Implements zero-cost retrieval augmentation through hybrid BM25 search with Reciprocal Rank Fusion and an O(1) Shadow Buffer fallback mechanism to prevent adversarial hallucinations.
主要结论
- D-MEM reduces API token consumption by 80% (from 1,648K to 319K tokens) while maintaining or improving accuracy on complex reasoning tasks under extreme noise conditions (ρ=0.75).
- The architecture achieves superior multi-hop reasoning performance (42.7% F1 vs. A-MEM's 27.0%, a +15.7 point gap) by preserving cleaner relational memory structures through selective cognitive restructuring.
- The Critic Router successfully gates 80% of computational resources while maintaining overall F1 score of 37.4% on standard benchmarks, demonstrating the effectiveness of the bio-inspired RPE mechanism.
摘要: The integration of structured, long-term memory is critical for the development of autonomous Large Language Model (LLM) agents. Recent advancements, such as the Agentic Memory (A-MEM) framework, have achieved significant progress by dynamically constructing and evolving knowledge graphs. However, existing architectures inherently operate as synchronous, "append-and-evolve-all" systems. Processing every user utterance through a computationally expensive O(N²) memory evolution pipeline introduces severe write-latency, unbounded API token costs, and catastrophic context window pollution caused by conversational noise. To address this scalability bottleneck, we introduce D-MEM (Dopamine-Gated Agentic Memory), a biologically inspired architecture that decouples short-term interaction from long-term cognitive restructuring. Drawing inspiration from the Dopamine-driven Reward Prediction Error (RPE) gating mechanism in the mammalian Ventral Tegmental Area (VTA), D-MEM implements a highly efficient Fast/Slow routing system. We introduce a lightweight Critic Router that continuously evaluates the Information Entropy (Surprise) and Long-term Utility of incoming stimuli. Routine inputs with low RPE are either bypassed entirely or cached in an O(1) fast-access buffer, preserving computational resources. Conversely, inputs generating a high RPE—such as factual contradictions or paradigm-shifting preference changes—trigger a "dopamine release" that activates the slow, O(N) deep memory evolution pipeline, actively reshaping the agent's global knowledge graph. To enable rigorous evaluation under realistic conditions, we further introduce the LoCoMo-Noise benchmark, which systematically injects controlled conversational noise into long-term dialogue sessions to simulate real-world interaction dynamics. Extensive evaluations demonstrate that D-MEM reduces API token consumption by over 80% and eliminates O(N²) write-latency bottlenecks, all while strictly outperforming synchronous baselines in complex multi-hop reasoning and adversarial resilience. By selectively gating cognitive restructuring and leveraging zero-cost retrieval augmentations, D-MEM provides a highly scalable and cost-efficient foundation for lifelong agentic memory. To support reproducibility, we open-source our implementation at https://github.com/london-and-tequila/dmem.