pith. machine review for the scientific record. sign in

arxiv: 2601.03236 · v2 · submitted 2026-01-06 · 💻 cs.AI

Recognition: unknown

MAGMA: A Multi-Graph based Agentic Memory Architecture for AI Agents

Authors on Pith no claims yet
classification 💻 cs.AI
keywords memorymagmareasoningagenticretrievalarchitecturecausalentity
0
0 comments X
read the original abstract

Memory-Augmented Generation (MAG) extends Large Language Models with external memory to support long-context reasoning, but existing approaches largely rely on semantic similarity over monolithic memory stores, entangling temporal, causal, and entity information. This design limits interpretability and alignment between query intent and retrieved evidence, leading to suboptimal reasoning accuracy. In this paper, we propose MAGMA, a multi-graph agentic memory architecture that represents each memory item across orthogonal semantic, temporal, causal, and entity graphs. MAGMA formulates retrieval as policy-guided traversal over these relational views, enabling query-adaptive selection and structured context construction. By decoupling memory representation from retrieval logic, MAGMA provides transparent reasoning paths and fine-grained control over retrieval. Experiments on LoCoMo and LongMemEval demonstrate that MAGMA consistently outperforms state-of-the-art agentic memory systems in long-horizon reasoning tasks.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 17 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. GroupMemBench: Benchmarking LLM Agent Memory in Multi-Party Conversations

    cs.CL 2026-05 conditional novelty 8.0

    GroupMemBench shows leading LLM memory systems reach only 46% average accuracy on multi-party tasks, with a simple BM25 baseline matching or beating most of them.

  2. Trojan Hippo: Weaponizing Agent Memory for Data Exfiltration

    cs.CR 2026-05 unverdicted novelty 8.0

    Trojan Hippo attacks on LLM agent memory achieve 85-100% success rates in data exfiltration across four memory backends even after 100 benign sessions, while evaluated defenses reduce success rates but impose varying ...

  3. Goal-Oriented Reasoning for RAG-based Memory in Conversational Agentic LLM Systems

    cs.AI 2026-05 unverdicted novelty 7.0

    Goal-Mem improves RAG memory retrieval in agentic LLMs by explicit goal decomposition and backward chaining via Natural Language Logic, outperforming nine baselines on multi-hop and implicit inference tasks.

  4. Learning Where to Embed: Noise-Aware Positional Embedding for Query Retrieval in Small-Object Detection

    cs.CV 2026-04 unverdicted novelty 7.0

    HELP uses heatmap-guided positional embeddings and a gradient mask to suppress background noise in queries, enabling efficient small-object detection with fewer decoder layers and parameters.

  5. Cognifold: Always-On Proactive Memory via Cognitive Folding

    cs.AI 2026-05 unverdicted novelty 6.0

    Cognifold is a new proactive memory architecture that folds event streams into emergent cognitive structures by extending complementary learning systems theory with a prefrontal intent layer and graph topology self-or...

  6. PRISM: Pareto-Efficient Retrieval over Intent-Aware Structured Memory for Long-Horizon Agents

    cs.CL 2026-05 unverdicted novelty 6.0

    PRISM achieves higher accuracy than baselines on long-horizon agent tasks at an order-of-magnitude smaller context budget by combining hierarchical bundle search, query-sensitive costing, evidence compression, and ada...

  7. HAGE: Harnessing Agentic Memory via RL-Driven Weighted Graph Evolution

    cs.AI 2026-05 unverdicted novelty 6.0

    HAGE proposes a trainable weighted graph memory framework with LLM intent classification, dynamic edge modulation, and RL optimization that improves long-horizon reasoning accuracy in agentic LLMs over static baselines.

  8. What Happens Inside Agent Memory? Circuit Analysis from Emergence to Diagnosis

    cs.AI 2026-05 unverdicted novelty 6.0

    In LLM agents, memory routing circuits emerge at 0.6B scale while content circuits appear only at 4B, and write/read operations recruit a pre-existing late-layer context hub instead of creating a new one, enabling a 7...

  9. What Happens Inside Agent Memory? Circuit Analysis from Emergence to Diagnosis

    cs.AI 2026-05 unverdicted novelty 6.0

    Circuit analysis reveals that routing circuits for agent memory emerge at 0.6B parameters while content circuits emerge at 4B, with a shared grounding hub and an unsupervised diagnostic achieving 76.2% accuracy for lo...

  10. SemanticAgent: A Semantics-Aware Framework for Text-to-SQL Data Synthesis

    cs.AI 2026-04 unverdicted novelty 6.0

    SemanticAgent introduces a three-stage semantic analysis, synthesis, and verification process that produces higher-quality text-to-SQL training data than prior execution-only methods.

  11. Self-Aware Vector Embeddings for Retrieval-Augmented Generation: A Neuroscience-Inspired Framework for Temporal, Confidence-Weighted, and Relational Knowledge

    cs.IR 2026-04 unverdicted novelty 6.0

    SmartVector augments embeddings with time, confidence, and relation signals plus a consolidation process, raising top-1 accuracy on versioned queries from 31% to 62% on a synthetic benchmark while cutting stale answer...

  12. Integrating Graphs, Large Language Models, and Agents: Reasoning and Retrieval

    cs.AI 2026-04 unverdicted novelty 6.0

    A structured survey organizing graph-LLM integration methods by purpose, modality, and strategy across application domains.

  13. Opal: Private Memory for Personal AI

    cs.CR 2026-04 unverdicted novelty 6.0

    Opal enables private long-term memory for personal AI by decoupling reasoning to a trusted enclave with a lightweight knowledge graph and piggybacking reindexing on ORAM accesses.

  14. RecoverFormer: End-to-End Contact-Aware Recovery for Humanoid Robots

    cs.RO 2026-04 unverdicted novelty 5.0

    A single causal-transformer policy with latent recovery modes and contact-affordance prediction enables humanoid robots to recover from 100-300 N pushes with 100% success in simulation, generalizing zero-shot across w...

  15. ActorMind: Emulating Human Actor Reasoning for Speech Role-Playing

    cs.SD 2026-04 unverdicted novelty 5.0

    ActorMind is a four-agent chain-of-thought framework that emulates human actors to produce spontaneous, emotion-infused speech responses for role-playing scenarios.

  16. FAST: A Synergistic Framework of Attention and State-space Models for Spatiotemporal Traffic Prediction

    cs.LG 2026-04 unverdicted novelty 4.0

    FAST uses a Temporal-Spatial-Temporal structure with attention and Mamba modules plus learnable embeddings to achieve better accuracy on traffic prediction tasks than previous models.

  17. Back to Basics: Let Conversational Agents Remember with Just Retrieval and Generation

    cs.CL 2026-04 unverdicted novelty 4.0

    A minimalist retrieval-and-generation framework using turn isolation and query-driven pruning outperforms complex memory systems by directly addressing signal sparsity and dual-level redundancy in dialogues.