Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      June 2, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      June 2, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      June 2, 2025

      How To Prevent WordPress SQL Injection Attacks

      June 2, 2025

      The Alters: Release date, mechanics, and everything else you need to know

      June 2, 2025

      I’ve fallen hard for Starsand Island, a promising anime-style life sim bringing Ghibli vibes to Xbox and PC later this year

      June 2, 2025

      This new official Xbox 4TB storage card costs almost as much as the Xbox SeriesXitself

      June 2, 2025

      I may have found the ultimate monitor for conferencing and productivity, but it has a few weaknesses

      June 2, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      May report 2025

      June 2, 2025
      Recent

      May report 2025

      June 2, 2025

      Write more reliable JavaScript with optional chaining

      June 2, 2025

      Deploying a Scalable Next.js App on Vercel – A Step-by-Step Guide

      June 2, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      The Alters: Release date, mechanics, and everything else you need to know

      June 2, 2025
      Recent

      The Alters: Release date, mechanics, and everything else you need to know

      June 2, 2025

      I’ve fallen hard for Starsand Island, a promising anime-style life sim bringing Ghibli vibes to Xbox and PC later this year

      June 2, 2025

      This new official Xbox 4TB storage card costs almost as much as the Xbox SeriesXitself

      June 2, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»Sakana AI Introduces Transformer²: A Machine Learning System that Dynamically Adjusts Its Weights for Various Tasks

    Sakana AI Introduces Transformer²: A Machine Learning System that Dynamically Adjusts Its Weights for Various Tasks

    January 17, 2025

    LLMs are essential in industries such as education, healthcare, and customer service, where natural language understanding plays a crucial role. Though highly versatile, LLMs’ challenge is adapting to new tasks. Most fine-tuning methods are resource and time-consuming. Moreover, the fine-tuning approach often results in overfitting or sacrificing general adaptability for task-specific performance. This is a barrier for LLMs to address dynamic new and unforeseen tasks and creates a bottleneck in the overall application.

    One of the most prominent methods to address these challenges is Low-Rank Adaptation (LoRA), which updates small, task-specific matrices while freezing the rest of the model’s parameters. Although this reduces the computational cost of fine-tuning, it has limitations, such as increased sensitivity to overfitting and the inability to scale efficiently across tasks. Moreover, LoRA’s design lacks inherent compositionality, limiting its ability to integrate multiple domain-specific skills.

    The researchers at Sakana AI and Institute of Science Tokyo introduced Transformer², a novel self-adaptive machine learning framework for large language models. Transformer² employs a groundbreaking method called Singular Value Fine-tuning (SVF), which adapts LLMs in real time to new tasks without extensive retraining. By focusing on selectively modifying the singular components of the model’s weight matrices, Transformer² enables dynamic task-specific adjustments. This innovation reduces the computational burden associated with fine-tuning, offering a scalable and efficient solution for self-adaptation.

    At the heart of Transformer² is the SVF method, which fine-tunes the singular values of weight matrices. This approach drastically minimizes the number of trainable parameters compared to traditional methods. Instead of altering the entire model, SVF leverages reinforcement learning to create compact “expert” vectors specialized for specific tasks. For the inference process, Transformer² works on a two-pass mechanism: the first is to analyze what the task might be and requires, and in the second, it dynamically integrates various relevant expert vectors to produce suitable behavior. Modularly, the approach ensures efficiency in addressing such a wide array of tasks through Transformer².

    Transformer² performed outstanding performance in extensive benchmark evaluations. For instance, the framework shows improvements of over 39% compared to baselines in visual question-answering domains. In mathematics-related problem-solving, when testing was done on the GSM8K datasets, this model showed its strength by winning more than any fine-tuning method, reaching about a 4% improvement in its performance. On programming tasks under the MBPP-pro benchmark, Transformer² displayed considerable accuracy improvements for domain-specific tasks and its general performance on various types of domains. As a result, Transformer² adapted efficiently to unseen tasks like ARC-Challenge and Humaneval by either maintaining or exceeding the baseline performance metrics.

    An important overall outcome was the SVF method’s efficiency. This improved training times and reduced the need for significant computational requirements as this method used fewer than 10% of the parameters required by LoRA. For example, for the GSM8K dataset, only 0.39 million parameters were needed for SVF training versus 6.82 million using LoRA to achieve higher performance. In addition, the model demonstrated good compositionality; vectors trained as an expert for one task could be reused and added together with others for a different, unrelated task, indicating the ability to scale up this Transformer² framework.

    Hostinger

    The researchers achieved this leap forward by addressing core limitations in existing methods, such as overfitting and inefficiency. By leveraging reinforcement learning, the SVF method provided principled regularization, preventing performance collapse on small datasets or narrow task domains. This allowed Transformer² to excel despite limited training data while maintaining task adaptability.

    Conclusion: A research team from Sakana AI provided a scalable and efficient solution to task-specific adaptation in LLMs. Transformer², with its SVF method, is a highly significant advancement within the field that will pave the way for computationally efficient self-adaptive AI systems that are highly versatile. This approach will answer present challenges and lay a foundation for future developments of adaptive AI technologies.


    Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 65k+ ML SubReddit.

    🚨 Recommend Open-Source Platform: Parlant is a framework that transforms how AI agents make decisions in customer-facing scenarios. (Promoted)

    The post Sakana AI Introduces Transformer²: A Machine Learning System that Dynamically Adjusts Its Weights for Various Tasks appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleMeet Tensor Product Attention (TPA): Revolutionizing Memory Efficiency in Language Models
    Next Article Top 30 AI Tools for Designers in 2025

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    June 2, 2025
    Machine Learning

    Off-Policy Reinforcement Learning RL with KL Divergence Yields Superior Reasoning in Large Language Models

    June 2, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Microsoft flags macOS bug — remotely bypassing Apple’s sophisticated System Integrity Protection (SIP) security solution and allowing unauthorized third-party rootkit installs

    News & Updates

    LLM for Biology: This Paper Discusses How Language Models can be Applied to Biological Research

    Development

    Microsoft enhances Data Wrangler with the ability to prepare data using natural language with new GitHub Copilot integration

    Development

    VideoDubber’s YouTube Channel ID Finder

    Web Development

    Highlights

    Machine Learning

    How Rufus doubled their inference speed and handled Prime Day traffic with AWS AI chips and parallel decoding

    May 28, 2025

    Large language models (LLMs) have revolutionized the way we interact with technology, but their widespread…

    A Shift From Browsers to Enterprise Targets: 2024 Zero-Day Exploitation Analysis

    May 1, 2025

    CVE-2025-44883 – D-Link FW-WGS-804HPT Stack Overflow Vulnerability

    May 20, 2025

    Understanding Accounts in Salesforce: A Detailed Guide

    June 27, 2024
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.