Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 13, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 13, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 13, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 13, 2025

      This $4 Steam Deck game includes the most-played classics from my childhood — and it will save you paper

      May 13, 2025

      Microsoft shares rare look at radical Windows 11 Start menu designs it explored before settling on the least interesting one of the bunch

      May 13, 2025

      NVIDIA’s new GPU driver adds DOOM: The Dark Ages support and improves DLSS in Microsoft Flight Simulator 2024

      May 13, 2025

      How to install and use Ollama to run AI LLMs on your Windows 11 PC

      May 13, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Community News: Latest PECL Releases (05.13.2025)

      May 13, 2025
      Recent

      Community News: Latest PECL Releases (05.13.2025)

      May 13, 2025

      How We Use Epic Branches. Without Breaking Our Flow.

      May 13, 2025

      I think the ergonomics of generators is growing on me.

      May 13, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      This $4 Steam Deck game includes the most-played classics from my childhood — and it will save you paper

      May 13, 2025
      Recent

      This $4 Steam Deck game includes the most-played classics from my childhood — and it will save you paper

      May 13, 2025

      Microsoft shares rare look at radical Windows 11 Start menu designs it explored before settling on the least interesting one of the bunch

      May 13, 2025

      NVIDIA’s new GPU driver adds DOOM: The Dark Ages support and improves DLSS in Microsoft Flight Simulator 2024

      May 13, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Artificial Intelligence»The future of AI training: DisTrO’s game-changing approach

    The future of AI training: DisTrO’s game-changing approach

    August 29, 2024

    Applied AI research group Nous Research developed an AI model training optimizer that could dramatically change the way AI models of the future will be trained.

    Traditionally, training an AI model requires massive data centers packed with GPUs like NVIDIA’s H100s, and high-speed interconnects to synchronize gradient and parameter updates between GPUs.

    Each training step requires vast amounts of data to be shared between thousands of GPUs. The required bandwidth means these GPUs need to be hardwired and physically close to each other. With DisTrO, Nous Research may have found a way to change that completely.

    As a model is trained, an optimizer algorithm adjusts the parameters of the model to minimize the loss function. The loss function measures the difference between the model’s predictions and the actual outcomes, and the goal is to reduce this loss as much as possible through iterative training.

    DisTrO-AdamW is a variation of the popular AdamW optimizer algorithm. DisTrO stands for “Distributed Training Over-the-Internet” and hints at what makes it so special.

    DisTrO-AdamW drastically reduces the amount of inter-GPU communication required during the training of large neural networks. And it does this without sacrificing the convergence rate or accuracy of the training process.

    In empirical tests, DisTrO-AdamW achieved an 857x reduction in inter-GPU communication. This means that the DisTrO approach can train models with comparable accuracy and speed but without the need for expensive, high-bandwidth hardware.

    For example, during the pre-training of a 1.2 billion LLM, DisTrO-AdamW matched the performance of traditional methods while reducing the required bandwidth from 74.4 GB to just 86.8 MB per training step.

    What if you could use all the computing power in the world to train a shared, open source AI model?

    Preliminary report: https://t.co/b1XgJylsnV

    Nous Research is proud to release a preliminary report on DisTrO (Distributed Training Over-the-Internet) a family of… pic.twitter.com/h2gQJ4m7lB

    — Nous Research (@NousResearch) August 26, 2024

    Implications for AI Training

    DisTrO’s impact on the AI landscape could be profound. By reducing the communication overhead, DisTrO allows for the decentralized training of large models. Instead of a data center with thousands of GPUs and high-speed switches, you could train a model on distributed commercial hardware connected via the internet.

    You could have a community of people contributing access to their computing hardware to train a model. Imagine millions of idle PCs or redundant Bitcoin mining rigs working together to train an open source model. DisTrO makes that possible, and there’s hardly any sacrifice in the time to train the model or its accuracy.

    Nous Research admits they’re not really sure why their approach works so well and more research is needed to see if it scales to larger models.

    If it does, training massive models might no longer be monopolized by Big Tech companies with the cash needed for large data centers. It could also have a big impact by reducing the environmental impact of energy and water-hungry data centers.

    The concept of decentralized training could also make some aspects of regulations like California’s proposed SB 1047 bill moot. The bill calls for additional safety checks for models that cost more than $100m to train.

    With DisTrO, a community of anonymous people with distributed hardware could create a ‘supercomputer’ of their own to train a model. It could also negate the US government’s efforts to stop China from importing NVIDIA’s most powerful GPUs.

    In a world where AI is becoming increasingly important, DisTrO offers a glimpse of a future where the development of these powerful tools is more inclusive, sustainable, and widespread.

    The post The future of AI training: DisTrO’s game-changing approach appeared first on DailyAI.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleBest LLM APIs for Document Data Extraction
    Next Article First AI + Education Summit is an international push for “AI fluency”

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 14, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-3623 – WordPress Uncanny Automator PHP Object Injection Vulnerability

    May 14, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    This foldable keyboard and mouse is a game-changer for work travel, and couldn’t be more portable

    News & Updates

    How managing networks differs on Windows 10 and Linux

    Development

    How to automate KendoUI dropdown with selenium webdriver and C#

    Development

    King Bob pleads guilty to Scattered Spider-linked cryptocurrency thefts from investors

    Development

    Highlights

    Email Automation with setTargetObjectId

    July 29, 2024

    As a Salesforce Developer, sending emails through Apex is a common task. One of the…

    AI Hallucinations: What Designers Need to Know

    February 10, 2025

    Where to buy MSI Claw 8 AI+ gaming handheld: Launch date and restock alerts

    March 24, 2025

    Distribution Release: EndeavourOS 2024.06.25

    June 30, 2024
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.