Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 27, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 27, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 27, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 27, 2025

      Don’t make this costly thermostat mistake – and the best place to put it

      May 27, 2025

      68% of tech vendor customer support to be handled by AI by 2028, says Cisco report

      May 27, 2025

      These $130 Anker earbuds have no business sounding this good for the price

      May 27, 2025

      Pocket is shutting down – here’s how to retrieve what little data you still can

      May 27, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Last Call: Early Access for NativePHP Ends This Week

      May 27, 2025
      Recent

      Last Call: Early Access for NativePHP Ends This Week

      May 27, 2025

      Setup Social Auth Redirects with Laravel Herd

      May 27, 2025

      Community News: Latest PECL Releases (05.27.2025)

      May 27, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft wants to make GamePad gaming faster on Chrome for Windows 11

      May 27, 2025
      Recent

      Microsoft wants to make GamePad gaming faster on Chrome for Windows 11

      May 27, 2025

      Windows 11 KB5058502 restores Win + C, direct download links for version 23H2

      May 27, 2025

      Leak hints at Windows 11’s new feature that optimizes performance, tied to Copilot branding (?)

      May 27, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Artificial Intelligence»New training approach could help AI agents perform better in uncertain conditions

    New training approach could help AI agents perform better in uncertain conditions

    January 29, 2025

    A home robot trained to perform household tasks in a factory may fail to effectively scrub the sink or take out the trash when deployed in a user’s kitchen, since this new environment differs from its training space.

    To avoid this, engineers often try to match the simulated training environment as closely as possible with the real world where the agent will be deployed.

    However, researchers from MIT and elsewhere have now found that, despite this conventional wisdom, sometimes training in a completely different environment yields a better-performing artificial intelligence agent.

    Their results indicate that, in some situations, training a simulated AI agent in a world with less uncertainty, or “noise,” enabled it to perform better than a competing AI agent trained in the same, noisy world they used to test both agents.

    The researchers call this unexpected phenomenon the indoor training effect.

    “If we learn to play tennis in an indoor environment where there is no noise, we might be able to more easily master different shots. Then, if we move to a noisier environment, like a windy tennis court, we could have a higher probability of playing tennis well than if we started learning in the windy environment,” explains Serena Bono, a research assistant in the MIT Media Lab and lead author of a paper on the indoor training effect.

    The researchers studied this phenomenon by training AI agents to play Atari games, which they modified by adding some unpredictability. They were surprised to find that the indoor training effect consistently occurred across Atari games and game variations.

    They hope these results fuel additional research toward developing better training methods for AI agents.

    “This is an entirely new axis to think about. Rather than trying to match the training and testing environments, we may be able to construct simulated environments where an AI agent learns even better,” adds co-author Spandan Madan, a graduate student at Harvard University.

    Bono and Madan are joined on the paper by Ishaan Grover, an MIT graduate student; Mao Yasueda, a graduate student at Yale University; Cynthia Breazeal, professor of media arts and sciences and leader of the Personal Robotics Group in the MIT Media Lab; Hanspeter Pfister, the An Wang Professor of Computer Science at Harvard; and Gabriel Kreiman, a professor at Harvard Medical School. The research will be presented at the Association for the Advancement of Artificial Intelligence Conference.

    Training troubles

    The researchers set out to explore why reinforcement learning agents tend to have such dismal performance when tested on environments that differ from their training space.

    Reinforcement learning is a trial-and-error method in which the agent explores a training space and learns to take actions that maximize its reward.

    The team developed a technique to explicitly add a certain amount of noise to one element of the reinforcement learning problem called the transition function. The transition function defines the probability an agent will move from one state to another, based on the action it chooses.

    Hostinger

    If the agent is playing Pac-Man, a transition function might define the probability that ghosts on the game board will move up, down, left, or right. In standard reinforcement learning, the AI would be trained and tested using the same transition function.

    The researchers added noise to the transition function with this conventional approach and, as expected, it hurt the agent’s Pac-Man performance.

    But when the researchers trained the agent with a noise-free Pac-Man game, then tested it in an environment where they injected noise into the transition function, it performed better than an agent trained on the noisy game.

    “The rule of thumb is that you should try to capture the deployment condition’s transition function as well as you can during training to get the most bang for your buck. We really tested this insight to death because we couldn’t believe it ourselves,” Madan says.

    Injecting varying amounts of noise into the transition function let the researchers test many environments, but it didn’t create realistic games. The more noise they injected into Pac-Man, the more likely ghosts would randomly teleport to different squares.

    To see if the indoor training effect occurred in normal Pac-Man games, they adjusted underlying probabilities so ghosts moved normally but were more likely to move up and down, rather than left and right. AI agents trained in noise-free environments still performed better in these realistic games.

    “It was not only due to the way we added noise to create ad hoc environments. This seems to be a property of the reinforcement learning problem. And that was even more surprising to see,” Bono says.

    Exploration explanations

    When the researchers dug deeper in search of an explanation, they saw some correlations in how the AI agents explore the training space.

    When both AI agents explore mostly the same areas, the agent trained in the non-noisy environment performs better, perhaps because it is easier for the agent to learn the rules of the game without the interference of noise.

    If their exploration patterns are different, then the agent trained in the noisy environment tends to perform better. This might occur because the agent needs to understand patterns it can’t learn in the noise-free environment.

    “If I only learn to play tennis with my forehand in the non-noisy environment, but then in the noisy one I have to also play with my backhand, I won’t play as well in the non-noisy environment,” Bono explains.

    In the future, the researchers hope to explore how the indoor training effect might occur in more complex reinforcement learning environments, or with other techniques like computer vision and natural language processing. They also want to build training environments designed to leverage the indoor training effect, which could help AI agents perform better in uncertain environments.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleGet Paid $48 EVERY 10 Minutes Using Deepseek and ChatGPT Operator: The 2025 Insane Money-Making Method (Step-by-Step Guide)
    Next Article These 5 MacOS video players for 4K are better than QuickTime – and most are free

    Related Posts

    Artificial Intelligence

    Understanding the faulty proteins linked to cancer and autism

    May 27, 2025
    Artificial Intelligence

    Fighting osteoporosis before it starts

    May 27, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    The Hype Around Signals

    Development

    First Nations Health Authority in Crisis: Cyberattack Shakes BC’s Healthcare Sector

    Development

    JavaScript Pulse: Weekly Dev Digest – May 2, 2025

    Web Development

    Does GitHub Copilot improve code quality? Here’s what the data says

    Development

    Highlights

    How publishers use Figma to help design the news

    February 20, 2025

    At trusted titles like The New York Times, The Economist, and The Minnesota Star Tribune,…

    Rilasciato Blender 4.4: Prestazioni Ottimizzate con Vulkan e Altre Novità

    March 19, 2025

    Exploring ShadCN: A Game-Changer for Component Libraries

    December 27, 2024

    Tap into Potential: Add Premium Touch to Your Raspberry Pi Projects With SunFounder’s 10-inch Display

    May 7, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.