Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      June 2, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      June 2, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      June 2, 2025

      How To Prevent WordPress SQL Injection Attacks

      June 2, 2025

      How Red Hat just quietly, radically transformed enterprise server Linux

      June 2, 2025

      OpenAI wants ChatGPT to be your ‘super assistant’ – what that means

      June 2, 2025

      The best Linux VPNs of 2025: Expert tested and reviewed

      June 2, 2025

      One of my favorite gaming PCs is 60% off right now

      June 2, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      `document.currentScript` is more useful than I thought.

      June 2, 2025
      Recent

      `document.currentScript` is more useful than I thought.

      June 2, 2025

      Adobe Sensei and GenAI in Practice for Enterprise CMS

      June 2, 2025

      Over The Air Updates for React Native Apps

      June 2, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      You can now open ChatGPT on Windows 11 with Win+C (if you change the Settings)

      June 2, 2025
      Recent

      You can now open ChatGPT on Windows 11 with Win+C (if you change the Settings)

      June 2, 2025

      Microsoft says Copilot can use location to change Outlook’s UI on Android

      June 2, 2025

      TempoMail — Command Line Temporary Email in Linux

      June 2, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»DeepSeek AI Introduces CODEI/O: A Novel Approach that Transforms Code-based Reasoning Patterns into Natural Language Formats to Enhance LLMs’ Reasoning Capabilities

    DeepSeek AI Introduces CODEI/O: A Novel Approach that Transforms Code-based Reasoning Patterns into Natural Language Formats to Enhance LLMs’ Reasoning Capabilities

    February 16, 2025

    Large Language Models (LLMs) have advanced significantly in natural language processing, yet reasoning remains a persistent challenge. While tasks such as mathematical problem-solving and code generation benefit from structured training data, broader reasoning tasks—like logical deduction, scientific inference, and symbolic reasoning—suffer from sparse and fragmented data. Traditional approaches, such as continual pretraining on code, often embed reasoning signals implicitly, making it difficult for models to generalize. Even text-to-code generation methods remain constrained by syntax-specific learning, limiting their applicability beyond programming-related tasks. A more structured approach is needed to expose LLMs to fundamental reasoning patterns while preserving logical rigor.

    DeepSeek AI Research presents CODEI/O, an approach that converts code-based reasoning into natural language. By transforming raw code into an input-output prediction format and expressing reasoning steps through Chain-of-Thought (CoT) rationales, CODEI/O allows LLMs to internalize core reasoning processes such as logic flow planning, decision tree traversal, and modular decomposition. Unlike conventional methods, CODEI/O separates reasoning from code syntax, enabling broader applicability while maintaining logical structure.

    Technical Overview and Benefits

    CODEI/O follows a structured data processing pipeline:

    1. Collecting Raw Code Files: Over 450K functions were gathered from multiple sources, including algorithm repositories and educational programming datasets.
    2. Standardizing the Data: The collected code was refined using DeepSeek-V2.5, ensuring clarity and execution compatibility.
    3. Generating Input-Output Pairs: Functions were executed with varying inputs to create structured training examples across diverse reasoning tasks.
    4. Generating Chain-of-Thought Reasoning: Using models like DeepSeek-V2.5, natural language explanations were generated to provide structured reasoning.
    5. Verification and Refinement: Predictions were validated through execution, with incorrect responses revised iteratively to improve reasoning accuracy.

    Key Features of CODEI/O:

    • Transformative Learning: Converts diverse code patterns into natural language CoT rationales, making reasoning transferable beyond programming contexts.
    • Syntax-Decoupled Learning: Separates logical reasoning from code syntax, improving adaptability across reasoning tasks.
    • Multi-Task Improvement: Enhances performance across symbolic, scientific, logical, mathematical, and commonsense reasoning domains.
    • Verifiability: Predictions can be validated through cached ground-truth matching or re-execution.
    • Iterative Refinement: A refined version, CODEI/O++, employs multi-turn revision to enhance reasoning accuracy.

    Empirical Results and Performance

    The impact of CODEI/O was tested across four base models (ranging from 7B to 30B parameters) on 14 reasoning benchmarks covering logic, symbolic inference, mathematics, scientific deduction, and commonsense reasoning.

    Findings:

    • Consistent Improvements: CODEI/O training led to higher scores across reasoning benchmarks compared to traditional pretraining methods.
    • Generalization Across Tasks: Unlike existing approaches that improve specific tasks but degrade performance elsewhere, CODEI/O showed balanced enhancements.
    • Comparison to Baselines: CODEI/O outperformed datasets such as OpenMathInstruct2, OpenCoder-SFT-Stage1, and WebInstruct.
    • Effectiveness of Multi-Turn Refinement: CODEI/O++ further improved results by iteratively refining incorrect responses, leveraging execution feedback for better reasoning quality.

    For instance, in logical and symbolic reasoning benchmarks such as BBH and CruxEval, CODEI/O led to notable performance gains. In math reasoning tasks (GSM8K, MATH, and MMLU-STEM), it demonstrated improvements over existing baselines. Even in commonsense reasoning, where code-based methods typically struggle, CODEI/O maintained robust results.

    Hostinger

    Conclusion

    CODEI/O presents a structured way to enhance LLMs’ reasoning by leveraging input-output transformations from real-world code. Instead of focusing on isolated reasoning tasks, it extracts universal reasoning patterns and translates them into natural language explanations. This structured learning approach ensures that models acquire robust reasoning skills across different domains.

    The introduction of multi-turn revision (CODEI/O++) further refines reasoning accuracy, demonstrating that iterative learning from execution feedback enhances model reliability. By making predictions verifiable, CODEI/O provides a scalable and reliable method for improving LLM reasoning.

    By bridging code-based and natural language reasoning, CODEI/O offers a promising direction for enhancing LLMs’ cognitive abilities beyond programming-related tasks.


    Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 75k+ ML SubReddit.

    🚨 Recommended Open-Source AI Platform: ‘IntellAgent is a An Open-Source Multi-Agent Framework to Evaluate Complex Conversational AI System’ (Promoted)

    The post DeepSeek AI Introduces CODEI/O: A Novel Approach that Transforms Code-based Reasoning Patterns into Natural Language Formats to Enhance LLMs’ Reasoning Capabilities appeared first on MarkTechPost.

    Source: Read More 

    Hostinger
    Facebook Twitter Reddit Email Copy Link
    Previous ArticleThis AI Paper from Apple Introduces a Distillation Scaling Law: A Compute-Optimal Approach for Training Efficient Language Models
    Next Article ReasonFlux: Elevating LLM Reasoning with Hierarchical Template Scaling

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    June 2, 2025
    Machine Learning

    MiMo-VL-7B: A Powerful Vision-Language Model to Enhance General Visual Understanding and Multimodal Reasoning

    June 2, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Enhancing Visual Search with Aesthetic Alignment: A Reinforcement Learning Approach Using Large Language Models and Benchmark Evaluations

    Development

    The best VPNs for Canada in 2025: Expert tested

    News & Updates

    The First Descendant: How to setup cross-play, invite friends, and enable cross-save

    Development

    CodeSOD: Brushing Up

    News & Updates

    Highlights

    CVE-2025-5389 – JeeWMS Improper Access Controls Remote Vulnerability

    May 31, 2025

    CVE ID : CVE-2025-5389

    Published : May 31, 2025, 7:15 p.m. | 1 hour, 43 minutes ago

    Description : A vulnerability, which was classified as critical, has been found in JeeWMS up to 20250504. Affected by this issue is the function dogenerateOne2Many of the file /generateController.do?dogenerateOne2Many of the component File Handler. The manipulation leads to improper access controls. The attack may be launched remotely. Continious delivery with rolling releases is used by this product. Therefore, no version details of affected nor updated releases are available.

    Severity: 6.3 | MEDIUM

    Visit the link for more details, such as CVSS details, affected products, timeline, and more…

    Icon Fonts vs. SVG

    July 8, 2024

    GOG hypes up support for a classic Capcom JRPG franchise returningtoPC after Dino Crisis: “We hear you loud and clear!”

    February 13, 2025

    Have a .gov email address? You can get Perplexity Pro free for a year – here’s now

    January 30, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.