INTELLECT-3: Scaling Agentic RL and MoE to SOTA Performance with prime-rl and 512 H200s
Description
Dive into the technical architecture and training pipeline behind INTELLECT-3, a 106B-parameter Mixture-of-Experts model (12B active) that achieves state-of-the-art performance for its size across math, code, science, and reasoning benchmarks, outperforming many larger frontier models.This episode provides an insider look into the large-scale reinforcement learning (RL) infrastructure stack developed by the Prime Intellect Team:
1. prime-rl Framework: Explore prime-rl, an open framework for large-scale asynchronous reinforcement learning tailored for agentic RL with first-class support for multi-turn interactions and tool use. Learn how its disaggregated architecture, leveraging FSDP 2 for the trainer and vLLM for inference, scales seamlessly to thousands of GPUs.
2. Training Efficiency: Discover critical optimizations for massive RL runs, including Continuous Batching and In-Flight Weight Updates, which are essential for maintaining high throughput and minimizing off-policyness, especially for long-context trajectories. Hear about how they achieved sequence lengths up to 72k using activation offloading.
3. MoE and Optimization: Understand the implementation details enabling efficient Mixture-of-Experts (MoE) training, the use of the Distributed Muon optimizer, and strategies for maintaining balanced expert load distribution.
4. Verifiable Environments: Examine the role of Verifiers and the Environments Hub in standardizing agentic RL training and evaluation, turning environments (including Math, Code, Deep Research, and Software Engineering) into reusable, versioned artifacts. We also detail the use of Prime Sandboxes for high-throughput, secure code execution needed for agentic coding environments.The sources confirm that the INTELLECT-3 model and the complete infrastructure stack, including the prime-rl framework and all environments, are open-source, aiming to narrow the gap between proprietary and open RL pipelines. The model was trained end-to-end on a 512 H200 cluster. This is a must-listen for ML practitioners building the next generation of reasoning and agentic models.





