Paper List
-
GOPHER: Optimization-based Phenotype Randomization for Genome-Wide Association Studies with Differential Privacy
This paper addresses the core challenge of balancing rigorous privacy protection with data utility when releasing full GWAS summary statistics, overco...
-
Real-time Cricket Sorting By Sex A low-cost embedded solution using YOLOv8 and Raspberry Pi
This paper addresses the critical bottleneck in industrial insect farming: the lack of automated, real-time sex sorting systems for Acheta domesticus ...
-
Training Dynamics of Learning 3D-Rotational Equivariance
This work addresses the core dilemma of whether to use computationally expensive equivariant architectures or faster symmetry-agnostic models with dat...
-
Fast and Accurate Node-Age Estimation Under Fossil Calibration Uncertainty Using the Adjusted Pairwise Likelihood
This paper addresses the dual challenge of computational inefficiency and sensitivity to fossil calibration errors in Bayesian divergence time estimat...
-
Few-shot Protein Fitness Prediction via In-context Learning and Test-time Training
This paper addresses the core challenge of accurately predicting protein fitness with only a handful of experimental observations, where data collecti...
-
scCluBench: Comprehensive Benchmarking of Clustering Algorithms for Single-Cell RNA Sequencing
This paper addresses the critical gap of fragmented and non-standardized benchmarking in single-cell RNA-seq clustering, which hinders objective compa...
-
Simulation and inference methods for non-Markovian stochastic biochemical reaction networks
This paper addresses the computational bottleneck of simulating and performing Bayesian inference for non-Markovian biochemical systems with history-d...
-
Assessment of Simulation-based Inference Methods for Stochastic Compartmental Models
This paper addresses the core challenge of performing accurate Bayesian parameter inference for stochastic epidemic models when the likelihood functio...
D-MEM: Dopamine-Gated Agentic Memory via Reward Prediction Error Routing
UC San Diego | Carnegie Mellon University
30秒速读
IN SHORT: This paper addresses the fundamental scalability bottleneck in LLM agentic memory systems: the O(N²) computational complexity and unbounded API token costs caused by processing every user utterance through expensive memory evolution pipelines, regardless of information value.
核心创新
- Methodology Introduces D-MEM, a bio-inspired architecture implementing dopamine-gated fast/slow routing based on Agentic Reward Prediction Error (RPE), decoupling short-term interaction from long-term cognitive restructuring.
- Methodology Develops the LoCoMo-Noise benchmark with systematic 75% noise injection (Filler: 40%, Status: 30%, Tangent: 30%) to simulate real-world conversational dynamics and evaluate memory robustness.
- Methodology Implements zero-cost retrieval augmentation through hybrid BM25 search with Reciprocal Rank Fusion and an O(1) Shadow Buffer fallback mechanism to prevent adversarial hallucinations.
主要结论
- D-MEM reduces API token consumption by 80% (from 1,648K to 319K tokens) while maintaining or improving accuracy on complex reasoning tasks under extreme noise conditions (ρ=0.75).
- The architecture achieves superior multi-hop reasoning performance (42.7% F1 vs. A-MEM's 27.0%, a +15.7 point gap) by preserving cleaner relational memory structures through selective cognitive restructuring.
- The Critic Router successfully gates 80% of computational resources while maintaining overall F1 score of 37.4% on standard benchmarks, demonstrating the effectiveness of the bio-inspired RPE mechanism.
摘要: The integration of structured, long-term memory is critical for the development of autonomous Large Language Model (LLM) agents. Recent advancements, such as the Agentic Memory (A-MEM) framework, have achieved significant progress by dynamically constructing and evolving knowledge graphs. However, existing architectures inherently operate as synchronous, "append-and-evolve-all" systems. Processing every user utterance through a computationally expensive O(N²) memory evolution pipeline introduces severe write-latency, unbounded API token costs, and catastrophic context window pollution caused by conversational noise. To address this scalability bottleneck, we introduce D-MEM (Dopamine-Gated Agentic Memory), a biologically inspired architecture that decouples short-term interaction from long-term cognitive restructuring. Drawing inspiration from the Dopamine-driven Reward Prediction Error (RPE) gating mechanism in the mammalian Ventral Tegmental Area (VTA), D-MEM implements a highly efficient Fast/Slow routing system. We introduce a lightweight Critic Router that continuously evaluates the Information Entropy (Surprise) and Long-term Utility of incoming stimuli. Routine inputs with low RPE are either bypassed entirely or cached in an O(1) fast-access buffer, preserving computational resources. Conversely, inputs generating a high RPE—such as factual contradictions or paradigm-shifting preference changes—trigger a "dopamine release" that activates the slow, O(N) deep memory evolution pipeline, actively reshaping the agent's global knowledge graph. To enable rigorous evaluation under realistic conditions, we further introduce the LoCoMo-Noise benchmark, which systematically injects controlled conversational noise into long-term dialogue sessions to simulate real-world interaction dynamics. Extensive evaluations demonstrate that D-MEM reduces API token consumption by over 80% and eliminates O(N²) write-latency bottlenecks, all while strictly outperforming synchronous baselines in complex multi-hop reasoning and adversarial resilience. By selectively gating cognitive restructuring and leveraging zero-cost retrieval augmentations, D-MEM provides a highly scalable and cost-efficient foundation for lifelong agentic memory. To support reproducibility, we open-source our implementation at https://github.com/london-and-tequila/dmem.