pith. machine review for the scientific record. sign in

arxiv: 2602.12036 · v2 · submitted 2026-02-12 · 💻 cs.CL

Recognition: unknown

Composition-RL: Compose Your Verifiable Prompts for Reinforcement Learning of Large Language Models

Authors on Pith no claims yet
classification 💻 cs.CL
keywords promptscomposition-rlverifiabletrainingbettercompositionaldataeffective
0
0 comments X
read the original abstract

Large-scale verifiable prompts underpin the success of Reinforcement Learning with Verifiable Rewards (RLVR), but they contain many uninformative examples and are costly to expand further. Recent studies focus on better exploiting limited training data by prioritizing hard prompts whose rollout pass rate is 0. However, easy prompts with a pass rate of 1 also become increasingly prevalent as training progresses, thereby reducing the effective data size. To mitigate this, we propose Composition-RL, a simple yet useful approach for better utilizing limited verifiable prompts targeting pass-rate-1 prompts. More specifically, Composition-RL automatically composes multiple problems into a new verifiable question and uses these compositional prompts for RL training. Extensive experiments across model sizes from 4B to 30B show that Composition-RL consistently improves reasoning capability over RL trained on the original dataset. Performance can be further boosted with a curriculum variant of Composition-RL that gradually increases compositional depth over training. Additionally, Composition-RL enables more effective cross-domain RL by composing prompts drawn from different domains. Codes, datasets, and models are available at https://github.com/XinXU-USTC/Composition-RL.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 2 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Rollout Pass-Rate Control: Steering Binary-Reward RL Toward Its Most Informative Regime

    cs.LG 2026-05 unverdicted novelty 6.0

    Prefix Sampling steers binary-reward agentic RL rollouts to a 50% pass rate to maximize learning signal, yielding up to 2.01x speedups on SWE-bench with maintained or improved verified performance.

  2. Rollout Pass-Rate Control: Steering Binary-Reward RL Toward Its Most Informative Regime

    cs.LG 2026-05 unverdicted novelty 6.0

    Prefix Sampling replays self-generated trajectory prefixes to control rollout pass rates to ~50% in binary-reward GRPO, delivering 2.01x and 1.55x speedups on Qwen3-14B/32B with slight score improvements on SWE-bench ...