Recognition: unknown
Experience Compression Spectrum: Unifying Memory, Skills, and Rules in LLM Agents
Pith reviewed 2026-05-10 08:41 UTC · model grok-4.3
The pith
The Experience Compression Spectrum unifies memory, skills, and rules in LLM agents along increasing compression levels and identifies the absence of adaptive cross-level compression as the missing diagonal.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
Mapping 20+ systems onto this spectrum reveals that every system operates at a fixed, predetermined compression level -- none supports adaptive cross-level compression, a gap we term the missing diagonal.
Load-bearing premise
That the compression ratios (5-20x, 50-500x, 1000x+) assigned to memory, skills, and rules are comparable across heterogeneous systems and that the low cross-community citation rate directly implies independent solving of shared sub-problems without solution exchange.
Figures
read the original abstract
As LLM agents scale to long-horizon, multi-session deployments, efficiently managing accumulated experience becomes a critical bottleneck. Agent memory systems and agent skill discovery both address this challenge -- extracting reusable knowledge from interaction traces -- yet a citation analysis of 1,136 references across 22 primary papers reveals a cross-community citation rate below 1%. We propose the \emph{Experience Compression Spectrum}, a unifying framework that positions memory, skills, and rules as points along a single axis of increasing compression (5--20$\times$ for episodic memory, 50--500$\times$ for procedural skills, 1,000$\times$+ for declarative rules), directly reducing context consumption, retrieval latency, and compute overhead. Mapping 20+ systems onto this spectrum reveals that every system operates at a fixed, predetermined compression level -- none supports adaptive cross-level compression, a gap we term the \emph{missing diagonal}. We further show that specialization alone is insufficient -- both communities independently solve shared sub-problems without exchanging solutions -- that evaluation methods are tightly coupled to compression levels, that transferability increases with compression at the cost of specificity, and that knowledge lifecycle management remains largely neglected. We articulate open problems and design principles for scalable, full-spectrum agent learning systems.
Editorial analysis
A structured set of objections, weighed in public.
Axiom & Free-Parameter Ledger
axioms (2)
- domain assumption Memory, skills, and rules can be ordered along a single axis of increasing compression
- domain assumption Low cross-citation rate indicates independent solution of shared sub-problems
invented entities (2)
-
Experience Compression Spectrum
no independent evidence
-
missing diagonal
no independent evidence
Forward citations
Cited by 1 Pith paper
-
Dynamic Skill Lifecycle Management for Agentic Reinforcement Learning
SLIM dynamically optimizes active external skills in agentic RL via leave-one-skill-out marginal contribution estimates and three lifecycle operations, outperforming baselines by 7.1% on ALFWorld and SearchQA while sh...
Reference graph
Works this paper leans on
-
[1]
Evoskill: Automated skill discovery for multi-agent systems.arXiv preprint arXiv:2603.02766,
Alzubi, S., Provenzano, N., Bingham, J., Chen, W., and Vu, T. EvoSkill: Automated skill discovery for multi-agent 6 Experience Compression Spectrum systems.arXiv preprint arXiv:2603.02766,
-
[2]
Constitutional AI: Harmlessness from AI Feedback
Bai, Y ., Kadavath, S., Kundu, S., Askell, A., Kernion, J., Jones, A., Chen, A., Goldie, A., Mirhoseini, A., McKin- non, C., et al. Constitutional AI: Harmlessness from AI feedback.arXiv preprint arXiv:2212.08073,
work page internal anchor Pith review Pith/arXiv arXiv
-
[3]
SEVerA: Verified Synthesis of Self-Evolving Agents
Banerjee, D., Xu, C., and Singh, G. SEVerA: Veri- fied synthesis of self-evolving agents.arXiv preprint arXiv:2603.25111,
work page internal anchor Pith review Pith/arXiv arXiv
-
[4]
Composer 2 technical report.arXiv preprint arXiv:2603.24477, 2026
Chan, A., Shalaby, A., Wettig, A., Sanger, A., Zhai, A., Ajay, A., Nair, A., Snell, C., Lu, C., Shen, C., Jia, E., Cassano, F., Liu, H., Chen, H., et al. Composer 2 technical report. arXiv preprint arXiv:2603.24477,
-
[5]
Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory
Chhikara, P., Khant, D., Aryan, S., Singh, T., and Yadav, D. Mem0: Building production-ready AI agents with scalable long-term memory.arXiv preprint arXiv:2504.19413,
work page internal anchor Pith review arXiv
-
[6]
Fang, J., Deng, X., Xu, H., Jiang, Z., Tang, Y ., Xu, Z., Deng, S., Yao, Y ., Wang, M., Qiao, S., Chen, H., and Zhang, N. LightMem: Lightweight and efficient memory- augmented generation.arXiv preprint arXiv:2510.18866,
-
[7]
Reinforced Self-Training (ReST) for Language Modeling
Gulcehre, C., Paine, T. L., Srinivasan, S., Konyushkova, K., Weerts, L., Sharma, A., Siddhant, A., Ahern, A., Wang, M., Gu, C., et al. Reinforced self-training (ReST) for language modeling.arXiv preprint arXiv:2308.08998,
-
[8]
Memory in the Age of AI Agents
Hu, Y ., Liu, S., Yue, Y ., Zhang, G., Liu, B., Zhu, F., Lin, J., Guo, H., Dou, S., Xi, Z., et al. Memory in the age of AI agents.arXiv preprint arXiv:2512.13564,
work page internal anchor Pith review arXiv
-
[9]
Cascade: Cumulative agentic skill creation through autonomous development and evolution
Huang, X., Chen, J., Fei, Y ., Li, Z., Schwaller, P., and Ceder, G. CASCADE: Cumulative agentic skill creation through autonomous development and evolution.arXiv preprint arXiv:2512.23880,
-
[10]
SoK: Agentic skills–beyond tool use in LLM agents.arXiv preprint arXiv:2602.20867, 2026
Jiang, Y ., Li, D., Deng, H., Ma, B., Wang, X., Wang, Q., and Yu, G. SoK: Agentic skills – beyond tool use in LLM agents.arXiv preprint arXiv:2602.20867,
-
[11]
Kang, J., Ji, M., Zhao, Z., and Bai, T. Memory OS of AI agent.arXiv preprint arXiv:2506.06326,
-
[12]
DSPy: Compiling Declarative Language Model Calls into Self-Improving Pipelines
Khattab, O., Singhvi, A., Maheshwari, P., Zhang, Z., San- thanam, K., Vardhamanan, S., Haq, S., Sharma, A., Joshi, T. T., Mober, H., et al. DSPy: Compiling declarative language model calls into self-improving pipelines.arXiv preprint arXiv:2310.03714,
work page internal anchor Pith review arXiv
-
[13]
Lam, C., Li, J., Zhang, L., and Zhao, K. Governing evolving memory in LLM agents: Risks, mechanisms, and the sta- bility and safety governed memory (SSGM) framework. arXiv preprint arXiv:2603.11768,
-
[14]
MemPO: Self-Memory Policy Optimization for Long-Horizon Agents
Li, R., Zhang, X., Yu, H., Duan, S., Li, X., Xiang, W., Liao, C., Guo, X., Li, Y ., and Suo, J. MemPO: Self-memory pol- icy optimization for long-horizon agents.arXiv preprint arXiv:2603.00680, 2026a. Li, X., Chen, W., Liu, Y ., Zheng, S., Chen, X., He, Y ., Li, Y ., You, B., Shen, H., Sun, J., et al. SkillsBench: Benchmarking how well agent skills work a...
work page internal anchor Pith review Pith/arXiv arXiv
-
[15]
arXiv preprint arXiv:2603.18718 , year=
Lin, M., Zhang, Z., Lu, H., Liu, H., Tang, X., He, Q., Zhang, X., and Wang, S. MemMA: Coordinating the memory cycle through multi-agent reasoning and in-situ self-evolution.arXiv preprint arXiv:2603.18718,
-
[16]
Trace2Skill: Distill Trajectory-Local Lessons into Transferable Agent Skills
Ni, J., Liu, Y ., Liu, X., Sun, Y ., Zhou, M., Cheng, P., Wang, D., Zhao, E., Jiang, X., and Jiang, G. Trace2Skill: Dis- till trajectory-local lessons into transferable agent skills. arXiv preprint arXiv:2603.25158,
work page internal anchor Pith review Pith/arXiv arXiv
-
[17]
Nie, A., Daull, X., Kuang, Z., Akkiraju, A., Chaudhuri, A., Piasevoli, M., Rong, R., Yuan, Y ., Choudhary, P., Xiao, S., Fakoor, R., Swaminathan, A., and Cheng, C.-A. Understanding the challenges in iterative generative opti- mization with LLMs.arXiv preprint arXiv:2603.23994,
-
[18]
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
7 Experience Compression Spectrum Shao, Z., Wang, P., Zhu, Q., Xu, R., Song, J., Zhang, M., Li, Y ., Wu, Y ., and Guo, D. DeepSeekMath: Pushing the limits of mathematical reasoning in open language models.arXiv preprint arXiv:2402.03300,
work page internal anchor Pith review Pith/arXiv arXiv
-
[19]
Voyager: An Open-Ended Embodied Agent with Large Language Models
Wang, G., Xie, Y ., Jiang, Y ., Mandlekar, A., Xiao, C., Zhu, Y ., Fan, L., and Anandkumar, A. V oyager: An open- ended embodied agent with large language models.arXiv preprint arXiv:2305.16291,
work page internal anchor Pith review arXiv
-
[20]
Wang, X., Liao, N., Wei, S., Tang, C., and Xiong, F. AutoAgent: Evolving cognition and elastic mem- ory orchestration for adaptive agents.arXiv preprint arXiv:2603.09716,
-
[21]
URL https://openreview.net/forum?id=ehfRiF0R3a
Wang, Y ., Takanobu, R., Liang, Z., Mao, Y ., Hu, Y ., McAuley, J., and Wu, X. Mem- α: Learning memory construction via reinforcement learning.arXiv preprint arXiv:2509.25911,
-
[22]
EvolveR: Self-Evolving LLM Agents through an Experience-Driven Lifecycle
Wu, R., Wang, X., Mei, J., Cai, P., Fu, D., Yang, C., Wen, L., Yang, X., Shen, Y ., Wang, Y ., and Shi, B. EvolveR: Self-evolving LLM agents through an experience-driven lifecycle.arXiv preprint arXiv:2510.16079,
work page internal anchor Pith review arXiv
-
[23]
SkillRL: Evolving Agents via Recursive Skill-Augmented Reinforcement Learning
Xia, P., Chen, J., Wang, H., Liu, J., Zeng, K., Wang, Y ., Han, S., Zhou, Y ., Zhao, X., Chen, H., Zheng, Z., Xie, C., and Yao, H. SkillRL: Evolving agents via recursive skill-augmented reinforcement learning.arXiv preprint arXiv:2602.08234,
work page internal anchor Pith review arXiv
-
[24]
Xiong, Y ., Hu, S., and Clune, J. Learning to continually learn via meta-learning agentic memory designs.arXiv preprint arXiv:2602.07755,
-
[25]
Agent Skills for Large Language Models: Architecture, Acquisition, Security, and the Path Forward
Xu, R. and Yan, Y . Agent skills for large language models: Architecture, acquisition, security, and the path forward. arXiv preprint arXiv:2602.12430,
work page internal anchor Pith review arXiv
-
[26]
A-MEM: Agentic Memory for LLM Agents
Xu, W., Liang, Z., Mei, K., Gao, H., Tan, J., and Zhang, Y . A-MEM: Agentic memory for LLM agents.arXiv preprint arXiv:2502.12110,
work page internal anchor Pith review arXiv
-
[27]
Yan, S., Yang, X., Huang, Z., Nie, E., Ding, Z., Li, Z., Ma, X., Bi, J., Kersting, K., Pan, J. Z., Sch ¨utze, H., Tresp, V ., and Ma, Y . Memory-R1: Enhancing large language model agents to manage and utilize memories via rein- forcement learning.arXiv preprint arXiv:2508.19828,
work page internal anchor Pith review arXiv
-
[28]
arXiv preprint arXiv:2602.05665 , year=
Yang, C., Zhou, C., Xiao, Y ., Dong, S., Zhuang, L., Zhang, Y ., Wang, Z., Hong, Z., Yuan, Z., Xiang, Z., et al. Graph- based agent memory: Taxonomy, techniques, and appli- cations.arXiv preprint arXiv:2602.05665, 2026a. Yang, Y ., Li, J., Pan, Q., Zhan, B., Cai, Y ., Du, L., Zhou, J., Chen, K., Chen, Q., Li, X., Zhang, B., and He, L. AutoSkill: Experienc...
-
[29]
MemSkill: Learning and Evolving Memory Skills for Self-Evolving Agents
Zhang, G., Geng, H., Yu, X., Yin, Z., Zhang, Z., Tan, Z., Zhou, H., Li, Z., Xue, X., Li, Y ., et al. The landscape of agentic reinforcement learning for LLMs: A survey. Transactions on Machine Learning Research (TMLR), 2026a. Zhang, H., Long, Q., Bao, J., Feng, T., Zhang, W., Yue, H., and Wang, W. MemSkill: Learning and evolving memory skills for self-evo...
work page internal anchor Pith review arXiv
-
[30]
SkillWeaver: Web Agents can Self-Improve by Discovering and Honing Skills
Zheng, B., Fatemi, M. Y ., Jin, X., Wang, Z. Z., Gandhi, A., Song, Y ., Gu, Y ., Srinivasa, J., Liu, G., Neubig, G., and Su, Y . SkillWeaver: Web agents can self-improve by discov- ering and honing skills.arXiv preprint arXiv:2504.07079,
work page internal anchor Pith review arXiv
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.