Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 16, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 16, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 16, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 16, 2025

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025

      Minecraft licensing robbed us of this controversial NFL schedule release video

      May 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      The power of generators

      May 16, 2025
      Recent

      The power of generators

      May 16, 2025

      Simplify Factory Associations with Laravel’s UseFactory Attribute

      May 16, 2025

      This Week in Laravel: React Native, PhpStorm Junie, and more

      May 16, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025
      Recent

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»This AI Paper Shows AI Model Collapses as Successive Model Generations Models are Recursively Trained on Synthetic Data

    This AI Paper Shows AI Model Collapses as Successive Model Generations Models are Recursively Trained on Synthetic Data

    July 30, 2024

    The phenomenon of “model collapse” presents a significant challenge in AI research, particularly for large language models (LLMs). When these models are trained on data that includes content generated by earlier versions of similar models, they tend to lose their ability to represent the true underlying data distribution over successive generations. This issue is critical because it compromises the performance and reliability of AI systems, which are increasingly integrated into various applications, from natural language processing to image generation. Addressing this challenge is essential to ensure that AI models can maintain their effectiveness and accuracy without degradation over time.

    Current methods to tackle training AI models involve using large datasets predominantly generated by humans. Techniques such as data augmentation, regularization, and transfer learning have been employed to enhance model robustness. However, these methods have limitations. For instance, they often require vast amounts of labeled data, which is not always feasible to obtain. Furthermore, existing models like variational autoencoders (VAEs) and Gaussian mixture models (GMMs) are susceptible to “catastrophic forgetting” and “data poisoning,” where the models either forget previously learned information or incorporate erroneous patterns from the data, respectively. These limitations hinder their performance, making them less suitable for applications requiring long-term learning and adaptation.

    The researchers present a novel approach involving a detailed examination of the “model collapse” phenomenon. They provide a theoretical framework and empirical evidence to demonstrate how models trained on recursively generated data gradually lose their ability to represent the true underlying data distribution. This approach specifically addresses the limitations of existing methods by highlighting the inevitability of model collapse in generative models, regardless of their architecture. The core innovation lies in identifying the sources of errors—statistical approximation error, functional expressivity error, and functional approximation error—that compound over generations, leading to model collapse. This understanding is crucial for developing strategies to mitigate such degradation, thereby offering a significant contribution to the field of AI.

    The technical approach employed in this research leverages datasets like wikitext2 to train language models, systematically illustrating the effects of model collapse through a series of controlled experiments. The researchers conducted detailed analyses of the perplexity of generated data points across multiple generations, revealing a significant increase in perplexity and indicating a clear degradation in model performance. Critical components of their methodology include Monte Carlo sampling and density estimation in Hilbert spaces, which provide a robust mathematical framework for understanding the propagation of errors across successive generations. These meticulously designed experiments also explore variations such as preserving a portion of the original training data to assess its impact on mitigating collapse.

    The findings demonstrate that models trained on recursively generated data exhibit a marked increase in perplexity, suggesting they become less accurate over time. Over successive generations, these models showed significant performance degradation, with higher perplexity and reduced variance in the generated data. The research also found that preserving a portion of the original human-generated data during training significantly mitigates the effects of model collapse, leading to better accuracy and stability in the models. The most notable result was the substantial improvement in accuracy when 10% of the original data was preserved, achieving an accuracy of 87.5% on a benchmark dataset, surpassing previous state-of-the-art results by 5%. This improvement highlights the importance of maintaining access to genuine human-generated data to sustain model performance.

    In conclusion, the research presents a comprehensive study on the phenomenon of model collapse, offering both theoretical insights and empirical evidence to highlight its inevitability in generative models. The proposed solution involves understanding and mitigating the sources of errors that lead to collapse. This work advances the field of AI by addressing a critical challenge that affects the long-term reliability of AI systems. By maintaining access to genuine human-generated data, the findings suggest, it is possible to sustain the benefits of training from large-scale data and prevent the degradation of AI models over successive generations.

    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

    Don’t Forget to join our 47k+ ML SubReddit

    Find Upcoming AI Webinars here

    The post This AI Paper Shows AI Model Collapses as Successive Model Generations Models are Recursively Trained on Synthetic Data appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleOptimizing Memory for Large-Scale NLP Models: A Look at MINI-SEQUENCE TRANSFORMER
    Next Article Lean Copilot: An AI Tool that Allows Large Language Models (LLMs) to be used in Lean for Proof Automation

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 16, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2022-4363 – Wholesale Market WooCommerce CSRF Vulnerability

    May 16, 2025
    Leave A Reply Cancel Reply

    Hostinger

    Continue Reading

    Radware Cloud Web App Firewall Vulnerability Let Attackers Bypass Filters

    Security

    Deskflow – keyboard and mouse sharing app

    Linux

    Creating a detailed Content Audit & Mapping Strategy for your next site build

    Development

    Highlights from Our Women in Digital Breakfast at Sitecore Symposium 2024

    Development
    Hostinger

    Highlights

    Development

    Turla Group Deploys LunarWeb and LunarMail Backdoors in Diplomatic Missions

    May 15, 2024

    An unnamed European Ministry of Foreign Affairs (MFA) and its three diplomatic missions in the…

    This AI Paper Introduces BitNet a4.8: A Highly Efficient and Accurate 4-bit LLM

    November 10, 2024

    Omen Gaming Hub Alternative: Top 5 Options to Try Right Now

    January 29, 2025

    CVE-2025-2238 – Vikinger WordPress Privilege Escalation Vulnerability

    April 25, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.