Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Representative Line: Brace Yourself

      September 18, 2025

      Beyond the Pilot: A Playbook for Enterprise-Scale Agentic AI

      September 18, 2025

      GitHub launches MCP Registry to provide central location for trusted servers

      September 18, 2025

      MongoDB brings Search and Vector Search to self-managed versions of database

      September 18, 2025

      Distribution Release: Security Onion 2.4.180

      September 18, 2025

      Distribution Release: Omarchy 3.0.1

      September 17, 2025

      Distribution Release: Mauna Linux 25

      September 16, 2025

      Distribution Release: SparkyLinux 2025.09

      September 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      AI Momentum and Perficient’s Inclusion in Analyst Reports – Highlights From 2025 So Far

      September 18, 2025
      Recent

      AI Momentum and Perficient’s Inclusion in Analyst Reports – Highlights From 2025 So Far

      September 18, 2025

      Shopping Portal using Python Django & MySQL

      September 17, 2025

      Perficient Earns Adobe’s Real-time CDP Specialization

      September 17, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Valve Survey Reveals Slight Retreat in Steam-on-Linux Share

      September 18, 2025
      Recent

      Valve Survey Reveals Slight Retreat in Steam-on-Linux Share

      September 18, 2025

      Review: Elecrow’s All-in-one Starter Kit for Pico 2

      September 18, 2025

      FOSS Weekly #25.38: GNOME 49 Release, KDE Drama, sudo vs sudo-rs, Local AI on Android and More Linux Stuff

      September 18, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»AREAL: Accelerating Large Reasoning Model Training with Fully Asynchronous Reinforcement Learning

    AREAL: Accelerating Large Reasoning Model Training with Fully Asynchronous Reinforcement Learning

    June 18, 2025

    Introduction: The Need for Efficient RL in LRMs

    Reinforcement Learning RL is increasingly used to enhance LLMs, especially for reasoning tasks. These models, known as Large Reasoning Models (LRMs), generate intermediate “thinking” steps before providing final answers, thereby improving performance on complex problems such as math and coding. However, training LRMs with RL at scale is challenging due to the need for massive parallelization and efficient system design. Current systems often rely on synchronous batch processing, where generation must wait for the longest output in a batch to finish, leading to GPU underutilization. Even newer methods still face bottlenecks, as they use outdated rollouts but remain batch-based.

    Background: Reinforcement Learning RL’s Impact on LLM Reasoning Abilities 

    Reinforcement Learning RL has become a widely used strategy to enhance the reasoning abilities of LLMs, particularly for tasks with clearly defined reward signals, such as math, coding, scientific reasoning, and tool use. These models typically improve by extending their chain-of-thought reasoning during training. Open-source efforts have shown that distilled, smaller models can also perform well on such tasks. Asynchronous RL methods, proven effective in gaming, have recently been explored for LLMs, but primarily in short-context or limited-overlap scenarios. Prior work has also investigated strategies, such as partial rollouts, to improve efficiency while maintaining training stability. 

    System Overview: Introducing AReaL

    Researchers from IIIS, Tsinghua University, Ant Research, and HKUST introduce AReaL, a fully asynchronous reinforcement learning system designed to train large reasoning models more efficiently. Unlike traditional synchronous systems, AReaL separates the generation and training processes; rollout workers continuously produce outputs, while training workers update the model in parallel as new data arrives. This design enhances GPU usage and accelerates training. To handle data staleness, AReaL utilizes a tailored version of PPO and incorporates optimizations such as dynamic batching and parallel reward services. On math and code tasks, AReaL achieves up to 2.77× faster training while maintaining or improving final model performance. 

    Technical Architecture: Key Components and Optimizations

    AREAL is designed to decouple generation and training across separate GPU clusters, improving scalability, hardware efficiency, and flexibility for reinforcement learning with large models. The system includes four main components: rollout workers that support interruptible generation and model updates, a reward service that evaluates responses, trainer workers that perform PPO updates, and a controller that coordinates the data flow. To address challenges such as data staleness and inconsistent policy versions, AREAL employs staleness-aware training and a decoupled PPO objective. Additionally, system-level optimizations such as pipelined CPU-GPU operations, non-blocking asynchronous requests, and dynamic sequence packing enhance training speed and GPU efficiency. 

    Experimental Results: Scaling and Performance

    AREAL was tested on math and coding tasks using distilled Qwen2 models of various sizes. It achieved 2–3 times faster training than prior methods, such as DeepScaleR and DeepCoder, while maintaining comparable accuracy. The system scales efficiently across GPUs and handles long context lengths (up to 32k tokens), outperforming synchronous methods’ key design features such as interruptible generation and dynamic microbatching, which boost training speed and hardware utilization. Algorithmically, AREAL’s decoupled PPO objective allows stable learning even with stale data, unlike standard PPO. Overall, AREAL balances speed and performance effectively, making it well-suited for large-scale RL training of language models. 

    Conclusion: Advancing Large-Scale RL for Language Models

    In conclusion, AREAL is an asynchronous reinforcement learning system designed to enhance the efficiency of training LLMs, particularly for tasks such as coding and mathematical reasoning. Unlike traditional synchronous methods that wait for all outputs before updating, AREAL allows generation and training to run in parallel. This decoupling reduces GPU idle time and boosts throughput. To ensure learning remains stable, AREAL introduces staleness-aware strategies and a modified PPO algorithm that effectively handles older training data. Experiments show that it delivers up to 2.77 times faster training than synchronous systems, without sacrificing accuracy, marking a step forward in scaling up RL for large models. 


    Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 100k+ ML SubReddit and Subscribe to our Newsletter.

    The post AREAL: Accelerating Large Reasoning Model Training with Fully Asynchronous Reinforcement Learning appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleHow Latent Vector Fields Reveal the Inner Workings of Neural Autoencoders
    Next Article AI-Powered Personalization: Redefining the Future of Customer Experience✨

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    September 3, 2025
    Machine Learning

    Announcing the new cluster creation experience for Amazon SageMaker HyperPod

    September 3, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    JMeter Tutorial: An End-to-End Guide

    Development

    Satya Nadella says Microsoft makes money every time you use ChatGPT: “Every day that ChatGPT succeeds is a fantastic day”

    News & Updates

    Fortnite lands itself the exclusive premiere of the latest Star Wars Disney+ show before you can watch it on TV

    News & Updates

    render – command line reader offering better readability of web pages

    Linux

    Highlights

    Development

    Making V8 eager to compile your JavaScript

    May 2, 2025

    #​734 — May 2, 2025 Read on the Web JavaScript Weekly GSAP v3.13: JavaScript Animation…

    CVE-2025-4145 – Netgear EX6200 Remote Buffer Overflow

    May 1, 2025

    CVE-2025-31198 – Apple macOS Symlink Path Handling Vulnerability

    May 29, 2025

    CVE-2023-45721 – HCL Leap Unauthenticated Directory Information Exposure

    April 30, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.