Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 16, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 16, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 16, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 16, 2025

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025

      Minecraft licensing robbed us of this controversial NFL schedule release video

      May 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      The power of generators

      May 16, 2025
      Recent

      The power of generators

      May 16, 2025

      Simplify Factory Associations with Laravel’s UseFactory Attribute

      May 16, 2025

      This Week in Laravel: React Native, PhpStorm Junie, and more

      May 16, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025
      Recent

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Researchers at FPT Software AI Center Introduce XMainframe: A State-of-the-Art Large Language Model (LLM) Specialized for Mainframe Modernization to Address the $100B Legacy Code Modernization

    Researchers at FPT Software AI Center Introduce XMainframe: A State-of-the-Art Large Language Model (LLM) Specialized for Mainframe Modernization to Address the $100B Legacy Code Modernization

    August 13, 2024

    Introduction

    Mainframe operating systems, originating in the 1940s, remain essential to critical sectors such as finance and government. However, the vast legacy of COBOL code—estimated by IBM to be around 200 to 220 billion lines—needs to be migrated to modern platforms and rewritten in contemporary programming languages. This task is monumental, with the cost of rewriting COBOL code using human resources estimated at 32 to 50 cents per line, presenting a $100 billion challenge. The time required for a complete rewrite by human programmers is still uncertain. These systems are often perceived as outdated, requiring significant maintenance and modernization. Addressing this challenge demands innovative tools capable of understanding and interacting with legacy codebases, a long-standing obstacle for the industry. The advent of Large Language Models (LLMs) offers a potential solution to this enduring problem. However, there are several concerns when applying LLMs to mainframe modernization.

    Challenges in Using LLMs for Mainframe Modernization:

    1. Limited Training on Mainframe Languages: While existing LLMs are trained on a wide range of languages, both natural and programming, they lack sufficient training on languages used in mainframes, such as COBOL. The relatively small amount of COBOL code available online leads to inadequate understanding and reasoning in these models.. Additionally, organizations tend to keep their mainframe codebases private due to the high security demands of financial-critical sectors, further limiting the available training data.

    2. Lack of Proper Benchmarks: The absence of comprehensive documentation and clear business goals for mainframe systems makes it difficult to develop benchmarks to evaluate the quality of LLMs in this domain. This hinders the ability to measure their effectiveness and reliability in mainframe modernization tasks.

    3. Complexity Beyond Code Generation: LLMs for coding are primarily trained for code generation, the most common use case in software engineering tasks. However, mainframe modernization involves more than just generating COBOL code—organizations aim to migrate their systems to other languages. Thus, LLMs must possess knowledge beyond code generation to effectively modernize these systems.

    XMainframe

    To address these challenges, researchers at FPT Software AI Center have developed XMainframe, a state-of-the-art large language model (LLM) specifically designed with expertise in mainframe legacy systems and COBOL codebases. The solution includes the creation of an extensive data collection pipeline to produce high-quality training datasets, significantly enhancing XMainframe’s performance in this specialized domain. Additionally, they introduce MainframeBench, a comprehensive benchmark for evaluating mainframe knowledge through multiple-choice questions, question answering, and COBOL code summarization. Empirical evaluations show that XMainframe consistently outperforms existing state-of-the-art LLMs in these tasks, achieving 30% higher accuracy than DeepSeek-Coder on multiple-choice questions, doubling the BLEU score of Mixtral-Instruct 8x7B on question-answering, and scoring six times higher than GPT-3.5 on COBOL summarization. This work underscores XMainframe’s potential to drive significant advancements in managing and modernizing legacy systems, ultimately enhancing productivity and saving time for software developers.

    Illustration of steps to collect data to build Mainframe:

    Results on MCQ:

    Results on Q&A

    Results on Code Summarization:

    Check out the Paper and GitHub. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

    Don’t Forget to join our 48k+ ML SubReddit

    Find Upcoming AI Webinars here

    Thanks to FPT Software AI Center for the thought leadership/ Resources for this article. FPT Software AI Center has supported us in this content/article.

    The post Researchers at FPT Software AI Center Introduce XMainframe: A State-of-the-Art Large Language Model (LLM) Specialized for Mainframe Modernization to Address the $100B Legacy Code Modernization appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleOutperforming Existing Models with Multi-Pass Refinement: This AI Paper from Amazon Unveils a New Era in Code Suggestion Tools
    Next Article HybridRAG: A Hybrid AI System Formed by Integrating Knowledge Graphs and Vector Retrieval Augmented Generation Outperforming both Individually

    Related Posts

    Machine Learning

    Salesforce AI Releases BLIP3-o: A Fully Open-Source Unified Multimodal Model Built with CLIP Embeddings and Flow Matching for Image Understanding and Generation

    May 16, 2025
    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 16, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    The best Windows tablets in 2025: Expert tested and reviewed

    News & Updates

    WEBRL: A Self-Evolving Online Curriculum Reinforcement Learning Framework for Training High-Performance Web Agents with Open LLMs

    Development

    LWiAI Podcast #206 – Llama 4, Nova Act, xAI buys X, PaperBench

    Artificial Intelligence

    Uni-MoE: A Unified Multimodal LLM based on Sparse MoE Architecture

    Development

    Highlights

    mlut – Atomic CSS toolkit with Sass and ergonomics for creating styles of any complexity

    August 14, 2024

    Comments Source: Read More 

    Complete Creative Constructions

    March 19, 2025

    Get a taste of Monster Hunter Wild’s mount system with Capcom’s best alternative to Pokémon

    June 13, 2024

    How to get started with Windows Recall on Windows 11

    December 20, 2024
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.