pith. machine review for the scientific record. sign in

Title resolution pending

6 Pith papers cite this work. Polarity classification is still indexing.

6 Pith papers citing it

fields

cs.LG 5 cs.SE 1

years

2026 6

representative citing papers

Interpretability Can Be Actionable

cs.LG · 2026-05-11 · conditional · novelty 6.0

Interpretability research should be judged by actionability—the degree to which its insights support concrete decisions and interventions—rather than explanatory power alone.

Flag Varieties: A Geometric Framework for Deep Network Alignment

cs.LG · 2026-05-11 · unverdicted · novelty 6.0

Alignment in deep networks is governed by flag varieties, with subspace intersection dimension as the unique reparameterization-invariant observable, explaining regularization and activation effects from first principles.

Self-Attention as Transport: Limits of Symmetric Spectral Diagnostics

cs.LG · 2026-05-06 · unverdicted · novelty 6.0

Symmetric spectral diagnostics on attention are structurally blind to flow direction, with asymmetry G as the sole control parameter, yielding a two-axis test that distinguishes bottleneck versus diffuse hallucination modes with opposite polarity.

citing papers explorer

Showing 6 of 6 citing papers.

  • Interpretability Can Be Actionable cs.LG · 2026-05-11 · conditional · none · ref 152

    Interpretability research should be judged by actionability—the degree to which its insights support concrete decisions and interventions—rather than explanatory power alone.

  • Flag Varieties: A Geometric Framework for Deep Network Alignment cs.LG · 2026-05-11 · unverdicted · none · ref 31

    Alignment in deep networks is governed by flag varieties, with subspace intersection dimension as the unique reparameterization-invariant observable, explaining regularization and activation effects from first principles.

  • Self-Attention as Transport: Limits of Symmetric Spectral Diagnostics cs.LG · 2026-05-06 · unverdicted · none · ref 35

    Symmetric spectral diagnostics on attention are structurally blind to flow direction, with asymmetry G as the sole control parameter, yielding a two-axis test that distinguishes bottleneck versus diffuse hallucination modes with opposite polarity.

  • Tracing the Thought of a Grandmaster-level Chess-Playing Transformer cs.LG · 2026-04-11 · unverdicted · none · ref 17

    Sparse replacement layers decompose the MLP and attention modules of a chess-playing transformer to reveal verifiable tactical reasoning pathways and parallel computation patterns.

  • Automated Attention Pattern Discovery at Scale in Large Language Models cs.LG · 2026-04-04 · unverdicted · none · ref 24

    AP-MAE reconstructs masked attention patterns in LLMs with high accuracy, generalizes across models, predicts generation correctness at 55-70%, and enables 13.6% accuracy gains via targeted interventions.

  • Carbon-Taxed Transformers: A Green Compression Pipeline for Overgrown Language Models cs.SE · 2026-04-28 · unverdicted · none · ref 60

    CTT is a compression pipeline for LLMs that achieves up to 49x memory reduction, 10x faster inference, 81% lower CO2 emissions, and retains 68-98% accuracy on code clone detection, summarization, and generation tasks.