pith. machine review for the scientific record. sign in

hub

arXiv preprint arXiv:2305.17493

10 Pith papers cite this work. Polarity classification is still indexing.

10 Pith papers citing it

hub tools

verdicts

UNVERDICTED 10

representative citing papers

Annotations Mitigate Post-Training Mode Collapse

cs.CL · 2026-05-11 · unverdicted · novelty 6.0

Annotation-anchored training reduces semantic diversity collapse in post-trained language models by a factor of six compared to standard supervised fine-tuning while preserving instruction-following and improving with scale.

Reinforced Self-Training (ReST) for Language Modeling

cs.CL · 2023-08-17 · unverdicted · novelty 6.0

ReST improves LLM translation quality on benchmarks via offline RL on self-generated data, achieving gains in a compute-efficient way compared to typical RLHF.

Textbooks Are All You Need

cs.CL · 2023-06-20 · unverdicted · novelty 6.0

A 1.3B-parameter code model trained on 7B tokens of curated textbook and synthetic data achieves 50.6% on HumanEval, indicating data quality can enable strong performance at small scale.

Position: No Retroactive Cure for Infringement during Training

cs.CR · 2026-04-20 · unverdicted · novelty 5.0

Post-hoc mitigation cannot retroactively cure infringement that occurred during unauthorized data ingestion and training because liability attaches to data lineage and retained expressive value in model weights.

citing papers explorer

Showing 10 of 10 citing papers.