⚖️CodeScaler: Scaling Code LLM Training and Test-Time Inference via Execution-Free Reward Models

Xiao Zhu1,*, Xinyu Zhou1,*, Boyu Zhu2,4, Hanxu Hu5, Mingzhe Du6, Haotian Zhang2, Huiming Wang2,†, Zhijiang Guo1,3,†
1LARK, HKUST(GZ) 2Kuaishou Technology 3HKUST 4UCL 5UZH 6NUS
* Equal Contribution, † Corresponding Author

News

🎉 [2024-02] We have released the CodeScaler paper, code, dataset and models!

Introduction

Reinforcement Learning from Verifiable Rewards (RLVR) has driven recent progress in code large language models by leveraging execution-based feedback from unit tests, but its scalability is fundamentally constrained by the availability and reliability of high-quality test cases. We propose CodeScaler, an execution-free reward model designed to scale both reinforcement learning training and test-time inference for code generation. CodeScaler is trained on carefully curated preference data derived from verified code problems and incorporates syntax-aware code extraction and validity-preserving reward shaping to ensure stable and robust optimization. Across five coding benchmarks, CodeScaler improves Qwen3-8B-Base by an average of +11.72 points, outperforming binary execution-based RL by +1.82 points, and enables scalable reinforcement learning on synthetic datasets without any test cases. At inference time, CodeScaler serves as an effective test-time scaling method, achieving performance comparable to unit test approaches while providing a 10× reduction in latency. Moreover, CodeScaler surpasses existing reward models on RM-Bench not only in the code domain (+3.3 points), but also in general and reasoning domains (+2.7 points on average).

Overview

CodeScaler Overview

We propose CodeScaler, an execution-free reward model designed to scale both reinforcement learning training and test-time inference for code generation. CodeScaler is trained on carefully curated preference data derived from verified code problems and incorporates syntax-aware code extraction and validity-preserving reward shaping to ensure stable and robust optimization.

Training Time Scaling (RL) Results

RL Results

We conduct reinforcement learning on Qwen3-8B-Base and Qwen3-14B-Base using the DeepCoder training set, which contains programming tasks with high-quality, verified test cases. We compare three sources of reward signals during RL: binary execution feedback and two other reward models. Our method consistently delivers the strongest gains. We attribute this to the dense and structured reward signals produced by our model. Unlike binary pass/fail feedback, our continuous scoring provides informative gradients by assigning intermediate rewards to partially correct or logically sound solutions. This enables the policy to better capture code structure, refine reasoning paths, and explore a broader solution space.

QueST Results

Training without test cases provides a key advantage, it enables us to scale the RL training data which in turn allows continuous performance improvements. Starting from Qwen3-8B-Base trained on DeepCoder with CodeScaler, we perform continual RL training on the synthetic problems dataset. The model exhibits steady and progressive performance gains as training proceeds with CodeScaler, demonstrating its scalability for large-scale RL training.

Test Time Scaling (BoN) Results

Test Time Scaling Results

By assigning scalar rewards to generated code, our model naturally induces a ranking over candidate solutions, making it directly applicable to Best-of-N sampling and efficient test-time scaling (TTS). We compare against unit-test-based methods, represented by CURE, which synthesizes unit tests for each problem and selects the solution that passes the most tests, as well as other reward models. CodeScaler substantially outperforms existing reward models and achieves performance comparable to CURE, while requiring 10× lower latency. This makes our method particularly attractive for real-world deployment, where both solution quality and inference efficiency are critical.

RM-Bench Results

Reward Model Benchmark Results

We evaluated CodeScaler on standard reward model benchmarks. Although CodeScaler was not designed as a general purpose reward model, testing it on these benchmarks still provides valuable insights. We used RM-Bench to compare CodeScaler against other reward models. Beyond the clear improvement in code tasks (73.6 → 76.9), we found a surprising result: even though CodeScaler was trained only on code, it also improved in general (Chat, 80.6 → 83.0) and reasoning domains (Math, 75.0 → 79.9). This indicates that learning to evaluate code sharpens the model’s overall judgment, even beyond the code domain.

Citation

If you find our work helpful, please consider citing:


          @misc{zhu2026codescalerscalingcodellm,
                title={CodeScaler: Scaling Code LLM Training and Test-Time Inference via Execution-Free Reward Models}, 
                author={Xiao Zhu and Xinyu Zhou and Boyu Zhu and Hanxu Hu and Mingzhe Du and Haotian Zhang and Huiming Wang and Zhijiang Guo},
                year={2026},
                eprint={2602.17684},
                archivePrefix={arXiv},
                primaryClass={cs.LG},
                url={https://arxiv.org/abs/2602.17684}, 
        }