Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      The Value-Driven AI Roadmap

      September 9, 2025

      This week in AI updates: Mistral’s new Le Chat features, ChatGPT updates, and more (September 5, 2025)

      September 6, 2025

      Designing For TV: Principles, Patterns And Practical Guidance (Part 2)

      September 5, 2025

      Neo4j introduces new graph architecture that allows operational and analytics workloads to be run together

      September 5, 2025

      Lenovo Legion Go 2 specs unveiled: The handheld gaming device to watch this October

      September 10, 2025

      As Windows 10 support ends, users weigh costly extended security program against upgrading to Windows 11

      September 10, 2025

      Lenovo’s Legion Glasses 2 update could change handheld gaming

      September 10, 2025

      Is Lenovo’s refreshed LOQ tower enough to compete? New OLED monitors raise the stakes at IFA 2025

      September 10, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      External Forces Reshaping Financial Services in 2025 and Beyond

      September 10, 2025
      Recent

      External Forces Reshaping Financial Services in 2025 and Beyond

      September 10, 2025

      Why It’s Time to Move from SharePoint On-Premises to SharePoint Online

      September 10, 2025

      Apple’s Big Move: The Future of Mobile

      September 10, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Lenovo Legion Go 2 specs unveiled: The handheld gaming device to watch this October

      September 10, 2025
      Recent

      Lenovo Legion Go 2 specs unveiled: The handheld gaming device to watch this October

      September 10, 2025

      As Windows 10 support ends, users weigh costly extended security program against upgrading to Windows 11

      September 10, 2025

      Lenovo’s Legion Glasses 2 update could change handheld gaming

      September 10, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»Unsupervised System 2 Thinking: The Next Leap in Machine Learning with Energy-Based Transformers

    Unsupervised System 2 Thinking: The Next Leap in Machine Learning with Energy-Based Transformers

    July 25, 2025

    Artificial intelligence research is rapidly evolving beyond pattern recognition and toward systems capable of complex, human-like reasoning. The latest breakthrough in this pursuit comes from the introduction of Energy-Based Transformers (EBTs)—a family of neural architectures specifically designed to enable “System 2 Thinking” in machines without relying on domain-specific supervision or restrictive training signals.

    From Pattern Matching to Deliberate Reasoning

    Human cognition is often described in terms of two systems: System 1 (fast, intuitive, automatic) and System 2 (slow, analytical, effortful). While today’s mainstream AI models excel at System 1 thinking—rapidly making predictions based on experience—most fall short on the deliberate, multi-step reasoning required for challenging or out-of-distribution tasks. Current efforts, such as reinforcement learning with verifiable rewards, are largely confined to domains where correctness is easy to check, like math or code, and struggle to generalize beyond them.

    Energy-Based Transformers: A Foundation for Unsupervised System 2 Thinking

    The key innovation of EBTs lies in their architectural design and training procedure. Instead of directly producing outputs in a single forward pass, EBTs learn an energy function that assigns a scalar value to each input-prediction pair, representing their compatibility or “unnormalized probability.” Reasoning, in turn, becomes an optimization process: starting from a random initial guess, the model iteratively refines its prediction through energy minimization—akin to how humans explore and check solutions before committing.

    This approach allows EBTs to exhibit three critical faculties for advanced reasoning, lacking in most current models:

    1. Dynamic Allocation of Computation: EBTs can devote more computational effort—more “thinking steps”—to harder problems or uncertain predictions as needed, instead of treating all tasks or tokens equally.
    2. Modeling Uncertainty Naturally: By tracking energy levels throughout the thinking process, EBTs can model their confidence (or lack thereof), particularly in complex, continuous domains like vision, where traditional models struggle.
    3. Explicit Verification: Each proposed prediction is accompanied by an energy score indicating how well it matches the context, enabling the model to self-verify and prefer answers it “knows” are plausible.

    Advantages Over Existing Approaches

    Unlike reinforcement learning or externally supervised verification, EBTs do not require hand-crafted rewards or extra supervision; their system 2 capabilities emerge directly from unsupervised learning objectives. Moreover, EBTs are inherently modality-agnostic—they scale across both discrete domains (like text and language) and continuous ones (such as images or video), a feat beyond the reach of most specialized architectures.

    Experimental evidence shows that EBTs not only improve downstream performance on language and vision tasks when allowed to “think longer,” but also scale more efficiently during training—in terms of data, compute, and model size—compared to state-of-the-art Transformer baselines. Notably, their ability to generalize improves as the task becomes more challenging or out-of-distribution, echoing findings in cognitive science about human reasoning under uncertainty.

    A Platform for Scalable Thinking and Generalization

    The Energy-Based Transformer paradigm signals a pathway toward more powerful and flexible AI systems, capable of adapting their reasoning depth to the demands of the problem. As data becomes a bottleneck for further scaling, EBTs’ efficiency and robust generalization can open doors to advances in modeling, planning, and decision-making across a wide array of domains.

    While current limitations remain—such as increased computational cost during training and challenges with highly multi-modal data distribution—future research is poised to build on the foundation laid by EBTs. Potential directions include combining EBTs with other neural paradigms, developing more efficient optimization strategies, and extending their application to new multimodal and sequential reasoning tasks.

    Summary

    Energy-Based Transformers represent a significant step towards machines that can “think” more like humans—not simply reacting reflexively, but pausing to analyze, verify, and adapt their reasoning for open-ended, complex problems across any modality.


    Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project.

    Meet the AI Dev Newsletter read by 40k+ Devs and Researchers from NVIDIA, OpenAI, DeepMind, Meta, Microsoft, JP Morgan Chase, Amgen, Aflac, Wells Fargo and 100s more [SUBSCRIBE NOW]

    The post Unsupervised System 2 Thinking: The Next Leap in Machine Learning with Energy-Based Transformers appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleDualDistill and Agentic-R1: How AI Combines Natural Language and Tool Use for Superior Math Problem Solving
    Next Article Simple Login Form Using HTML, CSS, and JavaScript

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    September 3, 2025
    Machine Learning

    Announcing the new cluster creation experience for Amazon SageMaker HyperPod

    September 3, 2025
    Leave A Reply Cancel Reply

    For security, use of Google's reCAPTCHA service is required which is subject to the Google Privacy Policy and Terms of Use.

    Continue Reading

    CVE-2025-4743 – Code-projects Employee Record System SQL Injection Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    “Crisp as a fresh chip” — this Alienware OLED monitor deal is fading fast, and you’ll regret missing it

    News & Updates

    كود خصم سكوات وولف 2025

    Web Development

    CVE-2025-3899 – Webserver Certificates Cross-site Scripting

    Common Vulnerabilities and Exposures (CVEs)

    Highlights

    Development

    How to Set Up Coolify in AWS EC2 and Have the Power to Do Anything in the Cloud

    July 1, 2025

    Coolify is an open-source, self-hostable platform that serves as an alternative to services like Heroku,…

    Hackers Can Now Exploit AI Models via PyTorch – Critical Bug Found

    April 21, 2025

    Windows 11 24H2’s Task Manager new CPU usage formula rolls out to everyone

    May 27, 2025

    Malicious Python Packages on PyPI Downloaded 39,000+ Times, Steal Sensitive Data

    April 5, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.