Paper List
-
Discovery of a Hematopoietic Manifold in scGPT Yields a Method for Extracting Performant Algorithms from Biological Foundation Model Internals
This work addresses the core challenge of extracting reusable, interpretable, and high-performance biological algorithms from the opaque internal repr...
-
MS2MetGAN: Latent-space adversarial training for metabolite–spectrum matching in MS/MS database search
This paper addresses the critical bottleneck in metabolite identification: the generation of high-quality negative training samples that are structura...
-
Toward Robust, Reproducible, and Widely Accessible Intracranial Language Brain-Computer Interfaces: A Comprehensive Review of Neural Mechanisms, Hardware, Algorithms, Evaluation, Clinical Pathways and Future Directions
This review addresses the core challenge of fragmented and heterogeneous evidence that hinders the clinical translation of intracranial language BCIs,...
-
Less Is More in Chemotherapy of Breast Cancer
通过纳入细胞周期时滞和竞争项,解决了现有肿瘤-免疫模型的过度简化问题,以定量比较化疗方案。
-
Fold-CP: A Context Parallelism Framework for Biomolecular Modeling
This paper addresses the critical bottleneck of GPU memory limitations that restrict AlphaFold 3-like models to processing only a few thousand residue...
-
Open Biomedical Knowledge Graphs at Scale: Construction, Federation, and AI Agent Access with Samyama Graph Database
This paper addresses the core pain point of fragmented biomedical data by constructing and federating large-scale, open knowledge graphs to enable sea...
-
Predictive Analytics for Foot Ulcers Using Time-Series Temperature and Pressure Data
This paper addresses the critical need for continuous, real-time monitoring of diabetic foot health by developing an unsupervised anomaly detection fr...
-
Hypothesis-Based Particle Detection for Accurate Nanoparticle Counting and Digital Diagnostics
This paper addresses the core challenge of achieving accurate, interpretable, and training-free nanoparticle counting in digital diagnostic assays, wh...
D-MEM: Dopamine-Gated Agentic Memory via Reward Prediction Error Routing
UC San Diego | Carnegie Mellon University
30秒速读
IN SHORT: This paper addresses the fundamental scalability bottleneck in LLM agentic memory systems: the O(N²) computational complexity and unbounded API token costs caused by processing every user utterance through expensive memory evolution pipelines, regardless of information value.
核心创新
- Methodology Introduces D-MEM, a bio-inspired architecture implementing dopamine-gated fast/slow routing based on Agentic Reward Prediction Error (RPE), decoupling short-term interaction from long-term cognitive restructuring.
- Methodology Develops the LoCoMo-Noise benchmark with systematic 75% noise injection (Filler: 40%, Status: 30%, Tangent: 30%) to simulate real-world conversational dynamics and evaluate memory robustness.
- Methodology Implements zero-cost retrieval augmentation through hybrid BM25 search with Reciprocal Rank Fusion and an O(1) Shadow Buffer fallback mechanism to prevent adversarial hallucinations.
主要结论
- D-MEM reduces API token consumption by 80% (from 1,648K to 319K tokens) while maintaining or improving accuracy on complex reasoning tasks under extreme noise conditions (ρ=0.75).
- The architecture achieves superior multi-hop reasoning performance (42.7% F1 vs. A-MEM's 27.0%, a +15.7 point gap) by preserving cleaner relational memory structures through selective cognitive restructuring.
- The Critic Router successfully gates 80% of computational resources while maintaining overall F1 score of 37.4% on standard benchmarks, demonstrating the effectiveness of the bio-inspired RPE mechanism.
摘要: The integration of structured, long-term memory is critical for the development of autonomous Large Language Model (LLM) agents. Recent advancements, such as the Agentic Memory (A-MEM) framework, have achieved significant progress by dynamically constructing and evolving knowledge graphs. However, existing architectures inherently operate as synchronous, "append-and-evolve-all" systems. Processing every user utterance through a computationally expensive O(N²) memory evolution pipeline introduces severe write-latency, unbounded API token costs, and catastrophic context window pollution caused by conversational noise. To address this scalability bottleneck, we introduce D-MEM (Dopamine-Gated Agentic Memory), a biologically inspired architecture that decouples short-term interaction from long-term cognitive restructuring. Drawing inspiration from the Dopamine-driven Reward Prediction Error (RPE) gating mechanism in the mammalian Ventral Tegmental Area (VTA), D-MEM implements a highly efficient Fast/Slow routing system. We introduce a lightweight Critic Router that continuously evaluates the Information Entropy (Surprise) and Long-term Utility of incoming stimuli. Routine inputs with low RPE are either bypassed entirely or cached in an O(1) fast-access buffer, preserving computational resources. Conversely, inputs generating a high RPE—such as factual contradictions or paradigm-shifting preference changes—trigger a "dopamine release" that activates the slow, O(N) deep memory evolution pipeline, actively reshaping the agent's global knowledge graph. To enable rigorous evaluation under realistic conditions, we further introduce the LoCoMo-Noise benchmark, which systematically injects controlled conversational noise into long-term dialogue sessions to simulate real-world interaction dynamics. Extensive evaluations demonstrate that D-MEM reduces API token consumption by over 80% and eliminates O(N²) write-latency bottlenecks, all while strictly outperforming synchronous baselines in complex multi-hop reasoning and adversarial resilience. By selectively gating cognitive restructuring and leveraging zero-cost retrieval augmentations, D-MEM provides a highly scalable and cost-efficient foundation for lifelong agentic memory. To support reproducibility, we open-source our implementation at https://github.com/london-and-tequila/dmem.