StructKV compresses LLM KV caches by tracking global in-degree centrality across network depth and dynamically selecting compression layers to preserve long-range dependencies better than local pruning methods.
GemFilter: Discovering Gems in Early Layers for Accelerated Long-Context LLMs,
2 Pith papers cite this work. Polarity classification is still indexing.
2
Pith papers citing it
years
2026 2verdicts
UNVERDICTED 2representative citing papers
A pre-execution size filter cuts repository tokens by 80-89% at sub-millisecond cost and raises file-level accuracy from 25% to 72% in a small CodeLlama evaluation.
citing papers explorer
-
StructKV: Preserving the Structural Skeleton for Scalable Long-Context Inference
StructKV compresses LLM KV caches by tracking global in-degree centrality across network depth and dynamically selecting compression layers to preserve long-range dependencies better than local pruning methods.
-
Correctness-Aware Repository Filtering Under Maximum Effective Context Window Constraints
A pre-execution size filter cuts repository tokens by 80-89% at sub-millisecond cost and raises file-level accuracy from 25% to 72% in a small CodeLlama evaluation.