Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Designing Better UX For Left-Handed People

      July 25, 2025

      This week in AI dev tools: Gemini 2.5 Flash-Lite, GitLab Duo Agent Platform beta, and more (July 25, 2025)

      July 25, 2025

      Tenable updates Vulnerability Priority Rating scoring method to flag fewer vulnerabilities as critical

      July 24, 2025

      Google adds updated workspace templates in Firebase Studio that leverage new Agent mode

      July 24, 2025

      Trump’s AI plan says a lot about open source – but here’s what it leaves out

      July 25, 2025

      Google’s new Search mode puts classic results back on top – how to access it

      July 25, 2025

      These AR swim goggles I tested have all the relevant metrics (and no subscription)

      July 25, 2025

      Google’s new AI tool Opal turns prompts into apps, no coding required

      July 25, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Laravel Scoped Route Binding for Nested Resource Management

      July 25, 2025
      Recent

      Laravel Scoped Route Binding for Nested Resource Management

      July 25, 2025

      Add Reactions Functionality to Your App With Laravel Reactions

      July 25, 2025

      saasykit/laravel-open-graphy

      July 25, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Sam Altman won’t trust ChatGPT with his “medical fate” unless a doctor is involved — “Maybe I’m a dinosaur here”

      July 25, 2025
      Recent

      Sam Altman won’t trust ChatGPT with his “medical fate” unless a doctor is involved — “Maybe I’m a dinosaur here”

      July 25, 2025

      “It deleted our production database without permission”: Bill Gates called it — coding is too complex to replace software engineers with AI

      July 25, 2025

      Top 6 new features and changes coming to Windows 11 in August 2025 — from AI agents to redesigned BSOD screens

      July 25, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»Unsupervised System 2 Thinking: The Next Leap in Machine Learning with Energy-Based Transformers

    Unsupervised System 2 Thinking: The Next Leap in Machine Learning with Energy-Based Transformers

    July 25, 2025

    Artificial intelligence research is rapidly evolving beyond pattern recognition and toward systems capable of complex, human-like reasoning. The latest breakthrough in this pursuit comes from the introduction of Energy-Based Transformers (EBTs)—a family of neural architectures specifically designed to enable “System 2 Thinking” in machines without relying on domain-specific supervision or restrictive training signals.

    From Pattern Matching to Deliberate Reasoning

    Human cognition is often described in terms of two systems: System 1 (fast, intuitive, automatic) and System 2 (slow, analytical, effortful). While today’s mainstream AI models excel at System 1 thinking—rapidly making predictions based on experience—most fall short on the deliberate, multi-step reasoning required for challenging or out-of-distribution tasks. Current efforts, such as reinforcement learning with verifiable rewards, are largely confined to domains where correctness is easy to check, like math or code, and struggle to generalize beyond them.

    Energy-Based Transformers: A Foundation for Unsupervised System 2 Thinking

    The key innovation of EBTs lies in their architectural design and training procedure. Instead of directly producing outputs in a single forward pass, EBTs learn an energy function that assigns a scalar value to each input-prediction pair, representing their compatibility or “unnormalized probability.” Reasoning, in turn, becomes an optimization process: starting from a random initial guess, the model iteratively refines its prediction through energy minimization—akin to how humans explore and check solutions before committing.

    This approach allows EBTs to exhibit three critical faculties for advanced reasoning, lacking in most current models:

    1. Dynamic Allocation of Computation: EBTs can devote more computational effort—more “thinking steps”—to harder problems or uncertain predictions as needed, instead of treating all tasks or tokens equally.
    2. Modeling Uncertainty Naturally: By tracking energy levels throughout the thinking process, EBTs can model their confidence (or lack thereof), particularly in complex, continuous domains like vision, where traditional models struggle.
    3. Explicit Verification: Each proposed prediction is accompanied by an energy score indicating how well it matches the context, enabling the model to self-verify and prefer answers it “knows” are plausible.

    Advantages Over Existing Approaches

    Unlike reinforcement learning or externally supervised verification, EBTs do not require hand-crafted rewards or extra supervision; their system 2 capabilities emerge directly from unsupervised learning objectives. Moreover, EBTs are inherently modality-agnostic—they scale across both discrete domains (like text and language) and continuous ones (such as images or video), a feat beyond the reach of most specialized architectures.

    Experimental evidence shows that EBTs not only improve downstream performance on language and vision tasks when allowed to “think longer,” but also scale more efficiently during training—in terms of data, compute, and model size—compared to state-of-the-art Transformer baselines. Notably, their ability to generalize improves as the task becomes more challenging or out-of-distribution, echoing findings in cognitive science about human reasoning under uncertainty.

    A Platform for Scalable Thinking and Generalization

    The Energy-Based Transformer paradigm signals a pathway toward more powerful and flexible AI systems, capable of adapting their reasoning depth to the demands of the problem. As data becomes a bottleneck for further scaling, EBTs’ efficiency and robust generalization can open doors to advances in modeling, planning, and decision-making across a wide array of domains.

    While current limitations remain—such as increased computational cost during training and challenges with highly multi-modal data distribution—future research is poised to build on the foundation laid by EBTs. Potential directions include combining EBTs with other neural paradigms, developing more efficient optimization strategies, and extending their application to new multimodal and sequential reasoning tasks.

    Summary

    Energy-Based Transformers represent a significant step towards machines that can “think” more like humans—not simply reacting reflexively, but pausing to analyze, verify, and adapt their reasoning for open-ended, complex problems across any modality.


    Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project.

    Meet the AI Dev Newsletter read by 40k+ Devs and Researchers from NVIDIA, OpenAI, DeepMind, Meta, Microsoft, JP Morgan Chase, Amgen, Aflac, Wells Fargo and 100s more [SUBSCRIBE NOW]

    The post Unsupervised System 2 Thinking: The Next Leap in Machine Learning with Energy-Based Transformers appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleDualDistill and Agentic-R1: How AI Combines Natural Language and Tool Use for Superior Math Problem Solving
    Next Article How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    July 25, 2025
    Machine Learning

    DualDistill and Agentic-R1: How AI Combines Natural Language and Tool Use for Superior Math Problem Solving

    July 25, 2025
    Leave A Reply Cancel Reply

    For security, use of Google's reCAPTCHA service is required which is subject to the Google Privacy Policy and Terms of Use.

    Continue Reading

    CVE-2025-49886 – WebGeniusLab Zikzag Core PHP RFI Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    Building A Drupal To Storyblok Migration Tool: An Engineering Perspective

    Tech & Work

    CVE-2025-7120 – Campcodes Complaint Management System SQL Injection

    Common Vulnerabilities and Exposures (CVEs)

    OpenFeign vs WebClient: How to Choose a REST Client for Your Spring Boot Project

    Development

    Highlights

    FCC clears Surface Laptop 13″, Pro 12″ with Snapdragon, rounded design

    May 4, 2025

    Based on three FCC (Federal Communications Commission) certifications spotted by Windows Latest, Microsoft plans to…

    Finding the AI Design Tool That Actually Works for Designers

    April 29, 2025

    CVE-2025-4980 – Netgear DGND3700 HTTP Information Disclosure Vulnerability

    May 20, 2025

    Xbox has become a Game Pass machine and nothing more — Is it enough to justify Microsoft’s console over a costly gaming PC?

    May 6, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.