HeLa-Mem: Hebbian Learning and Associative Memory for LLM Agents
Long-term memory is a critical challenge for Large Language Model agents, as fixed context windows cannot preserve coherence across extended interactions. Existing memory systems represent conversation history as unstructured embedding vectors, retrieving information through semantic similarity. This paradigm fails to capture the associative structure of human memory, wherein related experiences progressively strengthen interconnections through repeated co-activation. Inspired by cognitive neuroscience, we identify three mechanisms central to biological memory: association, consolidation, and spreading activation, which remain largely absent in current research. To bridge this gap, we propose HeLa-Mem, a bio-inspired memory architecture that models memory as a dynamic graph with Hebbian learning dynamics. HeLa-Mem employs a dual-level organization: (1) an episodic memory graph that evolves through co-activation patterns, and (2) a semantic memory store populated via Hebbian Distillation, wherein a Reflective Agent identifies densely connected memory hubs and distills them into structured, reusable semantic knowledge. This dual-path design leverages both semantic similarity and learned associations, mirroring the episodic-semantic distinction in human cognition. Experiments on LoCoMo demonstrate superior performance across four question categories while using significantly fewer context tokens. Code is available on GitHub: https://github.com/ReinerBRO/HeLa-Mem
Opening excerpt (first ~120 words) tap to expand
Computer Science > Computation and Language arXiv:2604.16839 (cs) [Submitted on 18 Apr 2026] Title:HeLa-Mem: Hebbian Learning and Associative Memory for LLM Agents Authors:Jinchang Zhu, Jindong Li, Cheng Zhang, Jiahong Liu, Menglin Yang View a PDF of the paper titled HeLa-Mem: Hebbian Learning and Associative Memory for LLM Agents, by Jinchang Zhu and 4 other authors View PDF HTML (experimental) Abstract:Long-term memory is a critical challenge for Large Language Model agents, as fixed context windows cannot preserve coherence across extended interactions. Existing memory systems represent conversation history as unstructured embedding vectors, retrieving information through semantic similarity.
…
Excerpt limited to ~120 words for fair-use compliance. The full article is at arXiv.org.