Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 16, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 16, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 16, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 16, 2025

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025

      Minecraft licensing robbed us of this controversial NFL schedule release video

      May 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      The power of generators

      May 16, 2025
      Recent

      The power of generators

      May 16, 2025

      Simplify Factory Associations with Laravel’s UseFactory Attribute

      May 16, 2025

      This Week in Laravel: React Native, PhpStorm Junie, and more

      May 16, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025
      Recent

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Kwai-STaR: An AI Framework that Transforms LLMs into State-Transition Reasoners to Improve Their Intuitive Reasoning Capabilities

    Kwai-STaR: An AI Framework that Transforms LLMs into State-Transition Reasoners to Improve Their Intuitive Reasoning Capabilities

    November 10, 2024

    Large Language Models find it challenging to understand Mathematical reasoning. Mathematical reasoning involves various cognitive tasks like understanding and manipulating mathematical concepts, solving problems, and making logical deductions. Existing methods in this domain have been established to enhance the mathematical ability of LLMs. However, few recognize the value of state transition for LLM reasoning, which can significantly improve the reasoning abilities of LLMs but has yet to be widely recognized or utilized.

    Current methods focus on enhancing the characteristic mathematical abilities of LLMs through training such as GPT, LLaMA, and MetaMath. These models use large-scale mathematical prompting to guide stepwise reasoning during problem-solving. CoT and Best-of-N explore how to fully harness the potential of LLMs during inference to boost mathematical performance. Monte Carlo Tree Search and Process Reward Model have achieved remarkable results by decomposing the problem-solving process into multiple steps while simultaneously providing timely rewards. However, these methods have limitations in efficiency and adaptability across different problem types.

    Kwai-STaR, a framework to transform general LLMs into state transition reasoners, which systematically solve problems by performing state transition, has been proposed to overcome this challenge.

    Researchers from Tsinghua University, Kuaishou Technology, the Institute of Automation, and the Chinese Academy of Sciences have proposed Kwai-STaR. The process involves three main steps: defining a state space for problem-solving, constructing a state-transition dataset, and training LLMs using a two-stage curriculum. The dataset contains two types of instances: a majority of correct cases and a minority of wrong-then-verified cases from the data generator and trained reasoner. The training strategy consists of two stages to maximize learning efficiency: a fundamental stage and an advanced stage. The fundamental stage trains the model with the majority of right cases, enabling it to solve relatively simple problems and to grasp the state-transition manner. The advanced stage includes pairs of wrong and verified cases to further strengthen the proficiency. Kwai-DStar is trained on benchmarks such as GSM8K, which showed Kwai-STaR’s impressive performance and efficiency. It also showed that Kwai-STaR achieves high accuracy rates with simpler inference processes than those required by traditional methods.

    In conclusion, Kwai-DStar transforms a traditional LLM into a state-transition reasoner, which enhances its reasoning capabilities for tackling mathematical problems. The current Kwai-STaR has only validated its effectiveness in the field of mathematics. While the mathematical domain is both challenging and representative, the potential of state space for enhancing LLM reasoning in general scenarios remains unverified, which limits the generalizability of the Kwai-STaR. Therefore, the researchers are actively working to provide additional experimental results in more diverse and general settings to demonstrate the generalizability of the Kwai-STaR approach further. 


    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 55k+ ML SubReddit.

    [AI Magazine/Report] Read Our Latest Report on ‘SMALL LANGUAGE MODELS‘

    The post Kwai-STaR: An AI Framework that Transforms LLMs into State-Transition Reasoners to Improve Their Intuitive Reasoning Capabilities appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleRAGCache: Optimizing Retrieval-Augmented Generation with Dynamic Caching
    Next Article This AI Paper Introduces BitNet a4.8: A Highly Efficient and Accurate 4-bit LLM

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 17, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-40906 – MongoDB BSON Serialization BSON::XS Multiple Vulnerabilities

    May 17, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Incorporate offline and online human – machine workflows into your generative AI applications on AWS

    Development

    Case Insensitive CSS Attribute Selector

    Development

    How does AI-driven DevOps Transform Software Development?

    Development

    This Xbox analysis chart shows how only 7% of games launched with “Play Anywhere” support in 2024 — a decline year-over-year

    News & Updates

    Highlights

    How Next.js Supercharges Load Times & Retains Customers

    February 24, 2025

    In today’s hyper-connected world, a website’s speed is no longer a mere technical detail —…

    Best NSFW Character AI Chat Online | Eros AI

    January 2, 2025

    How to Fix the UCMUCSI_FAILURE Error on Windows

    December 20, 2024

    Dozens of users can’t find the Robux gift cards in Microsoft Rewards anymore

    February 18, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.