{"total":8,"items":[{"citing_arxiv_id":"2605.11695","ref_index":14,"ref_count":1,"confidence":0.88,"is_internal_anchor":false,"paper_title":"Emergent Communication between Heterogeneous Visual Agents through Decentralized Learning","primary_cat":"cs.CV","submitted_at":"2026-05-12T07:51:56+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":6.0,"formal_verification":"none","one_line_summary":"Heterogeneous visual agents form shared symbols via decentralized Metropolis-Hastings captioning, where encoder similarity shapes the content and symmetry of the resulting language.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.05907","ref_index":31,"ref_count":1,"confidence":0.88,"is_internal_anchor":false,"paper_title":"Decoding Alignment without Encoding Alignment: A critique of similarity analysis in neuroscience","primary_cat":"q-bio.NC","submitted_at":"2026-05-07T09:17:47+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":6.0,"formal_verification":"none","one_line_summary":"Decoding alignment metrics can remain high and unchanged even when encoding manifold topology is causally altered, so they do not imply similar function or computation across neural populations.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.05851","ref_index":178,"ref_count":1,"confidence":0.88,"is_internal_anchor":false,"paper_title":"Hypothesis generation and updating in large language models","primary_cat":"cs.LG","submitted_at":"2026-05-07T08:24:54+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":6.0,"formal_verification":"none","one_line_summary":"LLMs exhibit Bayesian-like hypothesis updating with strong-sampling bias and an evaluation-generation gap but generalize poorly outside observed data.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.03885","ref_index":33,"ref_count":1,"confidence":0.88,"is_internal_anchor":false,"paper_title":"Raising the Ceiling: Better Empirical Fixation Densities for Saliency Benchmarking","primary_cat":"cs.CV","submitted_at":"2026-05-05T15:45:00+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":7.0,"formal_verification":"none","one_line_summary":"A mixture model with adaptive KDE and per-image cross-validation raises estimated human fixation consistency by 5-15% median log-likelihood and up to 2 AUC points over fixed-bandwidth Gaussian baselines.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.01374","ref_index":38,"ref_count":1,"confidence":0.88,"is_internal_anchor":false,"paper_title":"MTA: Multi-Granular Trajectory Alignment for Large Language Model Distillation","primary_cat":"cs.CL","submitted_at":"2026-05-02T10:52:49+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":7.0,"formal_verification":"none","one_line_summary":"MTA improves LLM knowledge distillation by aligning representations along layer-wise trajectories with adaptive granularity from words to phrases using dynamic structural and hidden representation alignment losses.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2605.01205","ref_index":38,"ref_count":1,"confidence":0.88,"is_internal_anchor":false,"paper_title":"SRA: Span Representation Alignment for Large Language Model Distillation","primary_cat":"cs.CL","submitted_at":"2026-05-02T02:44:12+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":5.0,"formal_verification":"none","one_line_summary":"SRA reframes cross-tokenizer LLM distillation as alignment of attention-weighted span centers of mass in a multi-particle dynamical system and reports consistent gains over prior CTKD baselines.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2604.21780","ref_index":36,"ref_count":1,"confidence":0.88,"is_internal_anchor":false,"paper_title":"Only Brains Align with Brains: Cross-Region Alignment Patterns Expose Limits of Normative Models","primary_cat":"q-bio.NC","submitted_at":"2026-04-23T15:38:11+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":6.0,"formal_verification":"none","one_line_summary":"Alignment pattern analysis reveals that models aligned to individual brain ROIs do not reproduce the stable cross-region alignment profiles observed across human subjects.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null},{"citing_arxiv_id":"2604.07886","ref_index":17,"ref_count":1,"confidence":0.88,"is_internal_anchor":false,"paper_title":"Linear Representations of Hierarchical Concepts in Language Models","primary_cat":"cs.CL","submitted_at":"2026-04-09T06:55:19+00:00","verdict":"UNVERDICTED","verdict_confidence":"LOW","novelty_score":6.0,"formal_verification":"none","one_line_summary":"Language models encode concept hierarchies as linear transformations that are domain-specific yet structurally similar across domains.","context_count":0,"top_context_role":null,"top_context_polarity":null,"context_text":null}],"limit":50,"offset":0}