pith. machine review for the scientific record. sign in

On-policy distillation of language models: Learning from self-generated mistakes

7 Pith papers cite this work. Polarity classification is still indexing.

7 Pith papers citing it

years

2026 7

clear filters

representative citing papers

Flow-OPD: On-Policy Distillation for Flow Matching Models

cs.CV · 2026-05-08 · unverdicted · novelty 6.0 · 2 refs

Flow-OPD applies on-policy distillation to flow matching models, achieving GenEval of 92 and OCR accuracy of 94 on Stable Diffusion 3.5 Medium while avoiding the seesaw effect of multi-reward optimization.

VISD: Enhancing Video Reasoning via Structured Self-Distillation

cs.CV · 2026-05-07 · unverdicted · novelty 5.0 · 2 refs

VISD adds structured privileged feedback from a judge model and a direction-magnitude decoupling trick to let VideoLLMs learn token-level credit assignment while keeping RL stable, yielding higher accuracy and roughly 2x faster convergence on video reasoning benchmarks.

citing papers explorer

Showing 1 of 1 citing paper after filters.