pith. machine review for the scientific record. sign in

arxiv: 2604.15583 · v2 · submitted 2026-04-16 · 💻 cs.DB

Recognition: unknown

SAGE: Selective Attention-Guided Extraction for Token-Efficient Document Indexing

Authors on Pith no claims yet

Pith reviewed 2026-05-10 08:28 UTC · model grok-4.3

classification 💻 cs.DB
keywords token reductionattention mechanismslong-context QAdocument indexingtraining-freeselective extractionRAG
0
0 comments X

The pith

SAGE extracts only the most relevant 10% of long documents using attention from a lightweight model, allowing accurate question answering with far fewer tokens.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Large language models struggle with the cost and slowness of processing entire long documents for answering questions. SAGE addresses this by using attention from a small local model to pick out only the question-relevant sections. It does this in one pass without training or tuning, then sends the shortened version to the main model. This lets users cut token counts by 90 percent while keeping answer quality competitive on standard benchmarks.

Core claim

SAGE performs a single prefilling pass with a lightweight LLM to generate a query-specific relevance heatmap from attention signals, applies differential attention strategies to isolate relevant evidence, selects the top-scoring document units within a user-defined token budget, and passes only this reduced context to a downstream LLM for answer generation, achieving competitive performance on long-document QA benchmarks at 10% context budget.

What carries the argument

The relevance heatmap derived from differential attention signals in a lightweight LLM's prefilling pass, which scores document units for question relevance to enable selective extraction.

If this is right

  • Users can reduce token usage by 90% for long document question answering tasks.
  • Performance remains competitive on benchmarks like QuALITY-hard without fine-tuning.
  • The method works plug-and-play across varied document structures and question types.
  • No complex calibration or task-specific tuning of retrievers is required.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Combining SAGE with other efficiency techniques could further optimize costs in production LLM systems.
  • This approach might lower risks of data leakage by not exposing full documents to the main model.
  • Testing on documents with different languages or formats could reveal broader applicability beyond English technical texts.

Load-bearing premise

Attention signals from a lightweight local LLM can reliably isolate question-relevant evidence that transfers to a different downstream LLM without task-specific tuning.

What would settle it

Running SAGE on a new long-document QA dataset where the 10% selected context leads to answers much worse than using the full document or a tuned RAG system would falsify the central claim.

Figures

Figures reproduced from arXiv: 2604.15583 by Chunwei Liu, Gerardo Vitagliano, Jun Chen, Matthew Russo, Michael Cafarella, Peter Baile Chen, Samuel Madden, Xinzhi Wang.

Figure 1
Figure 1. Figure 1: Heatmaps of ground-truth answer distributions [PITH_FULL_IMAGE:figures/full_fig_p002_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Overview of the SAGE pipeline. ○1 Attention computation utilizes chunked processing and length normalization to form a document-level score. ○2 Differential attention refines relevance by filtering structural noise. ○3 Unit-window selection extracts top text spans under the budget. ○4 Spans are assembled into a reduced-context prompt for final answer generation. document into chunks that fit within the con… view at source ↗
Figure 3
Figure 3. Figure 3: Effect of differential attention on token-level rele [PITH_FULL_IMAGE:figures/full_fig_p005_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Query normalization for differential attention. Each [PITH_FULL_IMAGE:figures/full_fig_p006_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Accuracy on QuALITY-hard versus average token [PITH_FULL_IMAGE:figures/full_fig_p009_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: LLM-judge accuracy on the Paper dataset. Accuracy vs. average token usage over documents for [PITH_FULL_IMAGE:figures/full_fig_p010_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Semantic similarity CDFs on the Paper dataset. CDFs of cosine similarity between generated and ground-truth answers, [PITH_FULL_IMAGE:figures/full_fig_p010_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Effect of differential attention on the Paper dataset. Semantic similarity CDFs compare Raw attention with differential [PITH_FULL_IMAGE:figures/full_fig_p011_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: LLM-judge accuracy on the Notice dataset. Accuracy versus average token usage for [PITH_FULL_IMAGE:figures/full_fig_p012_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: Semantic similarity CDFs on the Notice dataset. CDFs compare [PITH_FULL_IMAGE:figures/full_fig_p012_10.png] view at source ↗
read the original abstract

Large language models with long context windows can answer complex questions directly from full-length academic, technical, and policy documents, but passing entire documents is often costly, slow, and can degrade answer quality while increasing the risk of unnecessary data leakage. This paper targets the common setting of answering many heterogeneous questions over long document(s), where fixed position heuristics and standard retrieval-augmented generation (RAG) can fail due to document structure variability and weak query-chunk semantic similarity, which often requires task- and domain-specific tuning of embedding retrievers. We propose {Selective Attention-Guided Extraction} (\ourmethod), a training-free, plug-and-play context reduction framework that uses a lightweight local LLM to perform a single prefilling pass and convert language model attention signals into a query-specific relevance heatmap at configurable granularities. \ourmethod\ further introduces \emph{differential attention} strategies to better isolate question-relevant evidence, then selects the top-scoring units under a user-defined token budget and forwards only this reduced context to a downstream LLM for answer generation. \ourmethod\ surpasses traditional reduction techniques across multiple long-document QA benchmarks, notably securing a top-4 rank on QuALITY-hard while constrained to a 10\% context budget. This enables a 90\% reduction in tokens with competitive accuracy, without the need for model fine-tuning or complex calibration.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript proposes SAGE, a training-free, plug-and-play context reduction framework for long-document question answering. It performs a single prefilling pass with a lightweight local LLM to convert attention signals into a query-specific relevance heatmap at configurable granularities, applies differential attention strategies to isolate relevant evidence, selects the top-scoring units under a user-specified token budget, and forwards only the reduced context to a downstream LLM for answer generation. The central empirical claim is that SAGE surpasses traditional reduction techniques on multiple long-document QA benchmarks, including a top-4 rank on QuALITY-hard at a 10% context budget (90% token reduction) with competitive accuracy and no fine-tuning or calibration required.

Significance. If the reported benchmark results hold under rigorous controls, the work offers a practical, training-free alternative to RAG and position-based heuristics for token-efficient long-context inference. The use of off-the-shelf attention signals for query-specific extraction without task-specific tuning or embedding retriever calibration is a notable strength, as is the emphasis on heterogeneous documents where semantic similarity often fails. Reproducible attention-to-heatmap pipelines of this form could influence efficient deployment of LLMs on academic, technical, and policy corpora.

major comments (2)
  1. [§4] §4 (Experiments), Table 3 or equivalent results table: the top-4 ranking on QuALITY-hard at 10% budget is a load-bearing claim, yet the manuscript does not appear to report the precise ranking metric (e.g., exact-match F1, ROUGE, or human preference), the full list of competing methods, or variance across random seeds; without these, it is impossible to verify whether the result is robust or sensitive to baseline selection.
  2. [§3.2] §3.2 (Differential Attention): the transferability assumption—that attention heatmaps from the lightweight local LLM reliably identify evidence for a different downstream LLM across document structures—is central to the plug-and-play claim, but no ablation is described that isolates the effect of model mismatch (same vs. different model pairs) or document-type variability; this leaves the weakest assumption untested in the reported experiments.
minor comments (2)
  1. [Abstract] Abstract: the phrase 'top-4 rank' is used without defining the evaluation metric or the size of the comparison pool, which reduces clarity for readers unfamiliar with the QuALITY leaderboard.
  2. [§3.1] §3.1: the description of 'configurable granularities' for the relevance heatmap would benefit from an explicit example (e.g., token vs. sentence vs. paragraph units) and a short pseudocode snippet to make the pipeline reproducible from the text alone.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive and detailed feedback. We address each major comment point by point below, with planned revisions to strengthen the manuscript.

read point-by-point responses
  1. Referee: [§4] §4 (Experiments), Table 3 or equivalent results table: the top-4 ranking on QuALITY-hard at 10% budget is a load-bearing claim, yet the manuscript does not appear to report the precise ranking metric (e.g., exact-match F1, ROUGE, or human preference), the full list of competing methods, or variance across random seeds; without these, it is impossible to verify whether the result is robust or sensitive to baseline selection.

    Authors: We appreciate this clarification request. The QuALITY-hard benchmark uses accuracy as the standard evaluation metric for its multiple-choice questions, and the reported top-4 ranking is based on this metric at the 10% token budget. We will revise the manuscript to explicitly state the metric in the table caption and surrounding text. We will also expand the table to list all competing methods and their scores for full transparency. As SAGE is a deterministic, training-free approach with no learned parameters or random seeds in the selection process, variance across seeds does not apply; we will add a note to this effect. These changes will allow direct verification of the claim. revision: yes

  2. Referee: [§3.2] §3.2 (Differential Attention): the transferability assumption—that attention heatmaps from the lightweight local LLM reliably identify evidence for a different downstream LLM across document structures—is central to the plug-and-play claim, but no ablation is described that isolates the effect of model mismatch (same vs. different model pairs) or document-type variability; this leaves the weakest assumption untested in the reported experiments.

    Authors: We agree that an explicit test of the transferability assumption would strengthen the plug-and-play claims. Our experiments already pair a lightweight local LLM for attention extraction with larger, architecturally distinct downstream LLMs and evaluate across heterogeneous document types in multiple benchmarks. However, we did not include a dedicated ablation varying model pairs (matched vs. mismatched) or isolating document-type effects. In the revision, we will add a new ablation analysis comparing performance under same-model and cross-model configurations, along with breakdowns by document category where available in the benchmarks. revision: yes

Circularity Check

0 steps flagged

No significant circularity detected

full rationale

The provided abstract and description contain no mathematical derivations, equations, fitted parameters presented as predictions, or self-citation chains. The method is described as a training-free procedural pipeline that converts attention signals into a relevance heatmap and selects units under a token budget. No load-bearing step reduces by construction to its own inputs, and results are reported via external benchmarks rather than internal self-referential logic. This is the expected outcome for a purely empirical, non-derivational systems paper.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Abstract provides no equations, parameters, or postulates; method relies on existing LLM attention mechanisms but specifics of 'differential attention' and granularity choices are not detailed enough to classify.

pith-pipeline@v0.9.0 · 5562 in / 1183 out tokens · 38333 ms · 2026-05-10T08:28:40.664252+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

55 extracted references · 16 canonical work pages · 2 internal anchors

  1. [1]

    Amey Agrawal, Nitin Kedia, Anmol Agarwal, Jayashree Mohan, Nipun Kwa- tra, Souvik Kundu, Ramachandran Ramjee, and Alexey Tumanov. 2025. On Evaluating Performance of LLM Inference Serving Systems.arXiv preprint arXiv:2507.09019(2025)

  2. [2]

    Guoxuan Chen, Han Shi, Jiawei Li, Yihang Gao, Xiaozhe Ren, Yimeng Chen, Xin Jiang, Zhenguo Li, Weiyang Liu, and Chao Huang. 2024. Sepllm: Accelerate large language models by compressing one segment into one separator.arXiv preprint arXiv:2412.12094(2024)

  3. [3]

    Yilong Chen, Guoxia Wang, Junyuan Shang, Shiyao Cui, Zhenyu Zhang, Tingwen Liu, Shuohuan Wang, Yu Sun, Dianhai Yu, and Hua Wu. 2024. Nacl: A general and effective kv cache eviction framework for llm at inference time. InProceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 7913–7926

  4. [4]

    Xin Cheng, Xun Wang, Xingxing Zhang, Tao Ge, Si-Qing Chen, Furu Wei, Huishuai Zhang, and Dongyan Zhao. 2024. xrag: Extreme context compres- sion for retrieval-augmented generation with one token.Advances in Neural Information Processing Systems37 (2024), 109487–109516

  5. [5]

    Suyu Ge, Yunan Zhang, Liyuan Liu, Minjia Zhang, Jiawei Han, and Jianfeng Gao

  6. [6]

    Model tells you what to discard: Adaptive kv cache compression for llms.arXiv preprint arXiv:2310.01801,

    Model tells you what to discard: Adaptive kv cache compression for llms. arXiv preprint arXiv:2310.01801(2023)

  7. [7]

    Tao Ge, Jing Hu, Lei Wang, Xun Wang, Si-Qing Chen, and Furu Wei. 2023. In- context autoencoder for context compression in a large language model.arXiv preprint arXiv:2307.06945(2023)

  8. [8]

    Yuyao Ge, Shenghua Liu, Yiwei Wang, Lingrui Mei, Baolong Bi, Xuanshan Zhou, Jiayu Yao, Jiafeng Guo, and Xueqi Cheng. 2025. Focusing by Contrastive At- tention: Enhancing VLMs’ Visual Reasoning.arXiv preprint arXiv:2509.06461 (2025)

  9. [9]

    Saurabh Goyal, Anamitra Roy Choudhury, Saurabh Raje, Venkatesan Chakar- avarthy, Yogish Sabharwal, and Ashish Verma. 2020. Power-bert: Accelerating bert inference via progressive word-vector elimination. InInternational Confer- ence on Machine Learning. PMLR, 3690–3699

  10. [10]

    Ruishan Guo, Yibing Liu, Guoxin Ma, Yan Wang, Yueyang Zhang, Long Xia, Kecheng Chen, Zhiyuan Sun, and Daiting Shi. 2026. When Less is More: The LLM Scaling Paradox in Context Compression.arXiv preprint arXiv:2602.09789 (2026)

  11. [11]

    Huiqiang Jiang, Qianhui Wu, Chin-Yew Lin, Yuqing Yang, and Lili Qiu. 2023. Llmlingua: Compressing prompts for accelerated inference of large language models. InProceedings of the 2023 conference on empirical methods in natural language processing. 13358–13376

  12. [12]

    Huiqiang Jiang, Qianhui Wu, Xufang Luo, Dongsheng Li, Chin-Yew Lin, Yuqing Yang, and Lili Qiu. 2024. Longllmlingua: Accelerating and enhancing llms in long context scenarios via prompt compression. InProceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 1658–1677

  13. [13]

    Ziyan Jiang, Xueguang Ma, and Wenhu Chen. 2024. Longrag: Enhancing retrieval- augmented generation with long-context llms.arXiv preprint arXiv:2406.15319 (2024)

  14. [14]

    Bowen Jin, Jinsung Yoon, Jiawei Han, and Sercan O Arik. 2024. Long-context llms meet rag: Overcoming challenges for long inputs in rag.arXiv preprint arXiv:2410.05983(2024)

  15. [15]

    Jiajie Jin, Xiaoxi Li, Guanting Dong, Yuyao Zhang, Yutao Zhu, Yongkang Wu, Zhonghua Li, Ye Qi, and Zhicheng Dou. 2025. Hierarchical document refine- ment for long-context retrieval-augmented generation. InProceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 3502–3520

  16. [16]

    Hyukkyu Kang, Injung Kim, and Wook-Shin Han. 2025. TRIAL: Token Relations and Importance Aware Late-interaction for Accurate Text Retrieval. InProceed- ings of the 2025 Conference on Empirical Methods in Natural Language Processing. 16875–16888

  17. [17]

    Yannis Katsis, Saneem Chemmengath, Vishwajeet Kumar, Samarth Bharadwaj, Mustafa Canim, Michael Glass, Alfio Gliozzo, Feifei Pan, Jaydeep Sen, Karthik Sankaranarayanan, et al. 2022. Ait-qa: Question answering dataset over com- plex tables in the airline industry. InProceedings of the 2022 Conference of the North American Chapter of the Association for Comp...

  18. [18]

    Omar Khattab and Matei Zaharia. 2020. Colbert: Efficient and effective passage search via contextualized late interaction over bert. InProceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval. 39–48

  19. [19]

    Sehoon Kim, Sheng Shen, David Thorsley, Amir Gholami, Woosuk Kwon, Joseph Hassoun, and Kurt Keutzer. 2022. Learned token pruning for transformers. In Proceedings of the 28th ACM SIGKDD conference on knowledge discovery and data mining. 784–794

  20. [20]

    Xiang Lisa Li, Ari Holtzman, Daniel Fried, Percy Liang, Jason Eisner, Tatsunori B Hashimoto, Luke Zettlemoyer, and Mike Lewis. 2023. Contrastive decoding: Open-ended text generation as optimization. InProceedings of the 61st annual meeting of the association for computational linguistics (volume 1: Long papers). 12286–12312

  21. [21]

    Yuhong Li, Yingbing Huang, Bowen Yang, Bharat Venkitesh, Acyr Locatelli, Hanchen Ye, Tianle Cai, Patrick Lewis, and Deming Chen. 2024. Snapkv: Llm knows what you are looking for before generation.Advances in Neural Informa- tion Processing Systems37 (2024), 22947–22970

  22. [22]

    Zongqian Li, Yinhong Liu, Yixuan Su, and Nigel Collier. 2025. Prompt compres- sion for large language models: A survey. InProceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Lin- guistics: Human Language Technologies (Volume 1: Long Papers). 7182–7195

  23. [23]

    Junghwan Lim, Sungmin Lee, Dongseok Kim, Wai Ting Cheung, Beomgyu Kim, Taehwan Kim, Haesol Lee, Junhyeok Lee, Dongpin Oh, and Eunhwan Park. 2025. Grouped Differential Attention.arXiv preprint arXiv:2510.06949(2025)

  24. [24]

    Yiming Lin, Madelon Hulsebos, Ruiying Ma, Shreya Shankar, Sepanta Zeighami, Aditya G Parameswaran, and Eugene Wu. 2025. Querying templatized document collections with large language models. In2025 IEEE 41st International Conference on Data Engineering (ICDE). IEEE, 2422–2435

  25. [25]

    Barys Liskavets, Maxim Ushakov, Shuvendu Roy, Mark Klibanov, Ali Etemad, and Shane K Luke. 2025. Prompt compression with context-aware sentence encoding for fast and improved llm inference. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 39. 24595–24604

  26. [26]

    Chunwei Liu, Matthew Russo, Michael Cafarella, Lei Cao, Peter Baile Chen, Zui Chen, Michael Franklin, Tim Kraska, Samuel Madden, Rana Shahout, et al. [n.d.]. Palimpzest: Optimizing ai-powered analytics with declarative query processing

  27. [27]

    Di Liu, Meng Chen, Baotong Lu, Huiqiang Jiang, Zhenhua Han, Qianxi Zhang, Qi Chen, Chengruidong Zhang, Bailu Ding, Kai Zhang, et al. 2024. Retrievalattention: Accelerating long-context llm inference via vector retrieval.arXiv preprint arXiv:2409.10516(2024)

  28. [28]

    Jiaheng Liu, Dawei Zhu, Zhiqi Bai, Yancheng He, Huanxuan Liao, Haoran Que, Zekun Wang, Chenchen Zhang, Ge Zhang, Jiebin Zhang, et al. [n.d.]. A com- prehensive survey on long context language modeling. 2025.URL https://api. semanticscholar. org/CorpusID277271533 ([n. d.])

  29. [29]

    Nelson F Liu, Kevin Lin, John Hewitt, Ashwin Paranjape, Michele Bevilacqua, Fabio Petroni, and Percy Liang. 2024. Lost in the middle: How language models use long contexts.Transactions of the Association for Computational Linguistics 12 (2024), 157–173

  30. [30]

    Yubo Ma, Yuhang Zang, Liangyu Chen, Meiqi Chen, Yizhu Jiao, Xinze Li, Xinyuan Lu, Ziyu Liu, Yan Ma, Xiaoyi Dong, et al. 2024. Mmlongbench-doc: Benchmarking long-context document understanding with visualizations.Advances in Neural Information Processing Systems37 (2024), 95963–96010

  31. [31]

    mteb. [n.d.]. MTEB leaderboard - a hugging face space by mteb. https:// huggingface.co/spaces/mteb/leaderboard

  32. [32]

    Jesse Mu, Xiang Li, and Noah Goodman. 2024. Learning to compress prompts with gist tokens.Advances in Neural Information Processing Systems36 (2024)

  33. [33]

    Niklas Muennighoff, Nouamane Tazi, Loïc Magne, and Nils Reimers. 2023. Mteb: Massive text embedding benchmark. InProceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics. 2014–2037

  34. [34]

    Krista Opsahl-Ong, Michael J Ryan, Josh Purtell, David Broman, Christopher Potts, Matei Zaharia, and Omar Khattab. 2024. Optimizing instructions and demonstrations for multi-stage language model programs. InProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. 9340–9366

  35. [35]

    Richard Yuanzhe Pang, Alicia Parrish, Nitish Joshi, Nikita Nangia, Jason Phang, Angelica Chen, Vishakh Padmakumar, Johnny Ma, Jana Thompson, He He, and Samuel Bowman. [n.d.]. QuALITY Leaderboard — nyu-mll.github.io. https://nyu- mll.github.io/quality/. [Accessed 01-03-2026]

  36. [36]

    Richard Yuanzhe Pang, Alicia Parrish, Nitish Joshi, Nikita Nangia, Jason Phang, Angelica Chen, Vishakh Padmakumar, Johnny Ma, Jana Thompson, He He, and Samuel Bowman. 2022. QuALITY: Question Answering with Long Input Texts, Yes!. InProceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Lang...

  37. [37]

    Badri Patro and Vinay P Namboodiri. 2018. Differential attention for visual question answering. InProceedings of the IEEE conference on computer vision and pattern recognition. 7680–7688

  38. [38]

    Rui Qian, Xiaoyi Dong, Pan Zhang, Yuhang Zang, Shuangrui Ding, Dahua Lin, and Jiaqi Wang. 2024. Streaming long video understanding with large language models.Advances in Neural Information Processing Systems37 (2024), 119336– 119360

  39. [39]

    Keshav Santhanam, Omar Khattab, Jon Saad-Falcon, Christopher Potts, and Matei Zaharia. 2022. Colbertv2: Effective and efficient retrieval via lightweight late interaction. InProceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 3715–3734

  40. [40]

    Parth Sarthi, Salman Abdullah, Aditi Tuli, Shubh Khanna, Anna Goldie, and Christopher D Manning. 2024. Raptor: Recursive abstractive processing for tree-organized retrieval. InThe Twelfth International Conference on Learning Representations. 13

  41. [41]

    Kele Shao, TAO Keda, Kejia Zhang, Sicheng Feng, Mu Cai, Yuzhang Shang, Haoxuan You, Can Qin, Yang Sui, and Huan Wang. [n.d.]. A Survey of Token Compression for Efficient Multimodal Large Language Models.Transactions on Machine Learning Research([n. d.])

  42. [42]

    Hanrui Wang, Zhekai Zhang, and Song Han. 2021. Spatten: Efficient sparse attention architecture with cascade token and head pruning. In2021 IEEE inter- national symposium on high-performance computer architecture (HPCA). IEEE, 97–110

  43. [43]

    Yuetian Weng, Mingfei Han, Haoyu He, Xiaojun Chang, and Bohan Zhuang

  44. [44]

    In European Conference on Computer Vision

    Longvlm: Efficient long video understanding via large language models. In European Conference on Computer Vision. Springer, 453–470

  45. [45]

    David Wingate, Mohammad Shoeybi, and Taylor Sorensen. 2022. Prompt com- pression and contrastive conditioning for controllability and toxicity reduction in language models. InFindings of the Association for Computational Linguistics: EMNLP 2022. 5621–5634

  46. [46]

    Wenhao Wu, Yizhong Wang, Guangxuan Xiao, Hao Peng, and Yao Fu. 2024. Retrieval head mechanistically explains long-context factuality.arXiv preprint arXiv:2404.15574(2024)

  47. [47]

    Jingbo Yang, Bairu Hou, Wei Wei, Yujia Bao, and Shiyu Chang. 2025. Kvlink: Accelerating large language models via efficient kv cache reuse.arXiv preprint arXiv:2502.16002(2025)

  48. [48]

    Deming Ye, Yankai Lin, Yufei Huang, and Maosong Sun. 2021. Tr-bert: Dy- namic token reduction for accelerating bert inference. InProceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 5798–5809

  49. [49]

    Tianzhu Ye, Li Dong, Yuqing Xia, Yutao Sun, Yi Zhu, Gao Huang, and Furu Wei

  50. [50]

    Differential transformer.arXiv preprint arXiv:2410.05258(2024)

  51. [51]

    Tan Yu, Anbang Xu, and Rama Akkiraju. 2024. In defense of rag in the era of long-context language models.arXiv preprint arXiv:2409.01666(2024)

  52. [52]

    Peiyuan Zhang, Kaichen Zhang, Bo Li, Guangtao Zeng, Jingkang Yang, Yuanhan Zhang, Ziyue Wang, Haoran Tan, Chunyuan Li, and Ziwei Liu. 2024. Long context transfer from language to vision.arXiv preprint arXiv:2406.16852(2024)

  53. [53]

    Zhenyu Zhang, Runjin Chen, Shiwei Liu, Zhewei Yao, Olatunji Ruwase, Beidi Chen, Xiaoxia Wu, and Zhangyang Wang. 2024. Found in the middle: How language models use long contexts better via plug-and-play positional encoding. Advances in Neural Information Processing Systems37 (2024), 60755–60775

  54. [54]

    Zhenyu Zhang, Ying Sheng, Tianyi Zhou, Tianlong Chen, Lianmin Zheng, Ruisi Cai, Zhao Song, Yuandong Tian, Christopher Ré, Clark Barrett, et al. 2024. H2o: Heavy-hitter oracle for efficient generative inference of large language models. Advances in Neural Information Processing Systems36 (2024)

  55. [55]

    Xiabin Zhou, Wenbin Wang, Minyan Zeng, Jiaxian Guo, Xuebo Liu, Li Shen, Min Zhang, and Liang Ding. 2024. DynamicKV: Task-aware adaptive KV cache compression for long context LLMs.arXiv preprint arXiv:2412.14838(2024). 14