Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 18, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 18, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 18, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 18, 2025

      I need to see more from Lenovo’s most affordable gaming desktop, because this isn’t good enough

      May 18, 2025

      Gears of War: Reloaded — Release date, price, and everything you need to know

      May 18, 2025

      I’ve been using the Logitech MX Master 3S’ gaming-influenced alternative, and it could be your next mouse

      May 18, 2025

      Your Android devices are getting several upgrades for free – including a big one for Auto

      May 18, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      YTConverter™ lets you download YouTube videos/audio cleanly via terminal — especially great for Termux users.

      May 18, 2025
      Recent

      YTConverter™ lets you download YouTube videos/audio cleanly via terminal — especially great for Termux users.

      May 18, 2025

      NodeSource N|Solid Runtime Release – May 2025: Performance, Stability & the Final Update for v18

      May 17, 2025

      Big Changes at Meteor Software: Our Next Chapter

      May 17, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      I need to see more from Lenovo’s most affordable gaming desktop, because this isn’t good enough

      May 18, 2025
      Recent

      I need to see more from Lenovo’s most affordable gaming desktop, because this isn’t good enough

      May 18, 2025

      Gears of War: Reloaded — Release date, price, and everything you need to know

      May 18, 2025

      I’ve been using the Logitech MX Master 3S’ gaming-influenced alternative, and it could be your next mouse

      May 18, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Researchers at Stanford Introduce TEXTGRAD: A Powerful AI Framework Performing Automatic “Differentiation” via Text

    Researchers at Stanford Introduce TEXTGRAD: A Powerful AI Framework Performing Automatic “Differentiation” via Text

    June 13, 2024

    Artificial intelligence (AI) is experiencing a paradigm shift, with breakthroughs driven by systems orchestrating multiple large language models (LLMs) and other complex components. This progression has highlighted the need for effective optimization methods for these compound AI systems, where automatic differentiation comes into play. Automatic differentiation has revolutionized the training of neural networks, and now researchers seek to apply similar principles to optimize more complex AI systems via textual feedback from LLMs.

    One significant challenge in AI is optimizing compound systems that involve multiple components, such as LLMs, simulators, and web search tools. Traditional methods rely heavily on experts’ manual adjustments, which are time-consuming and prone to human error. Therefore, there is a pressing need for principled and automated optimization methods that can handle the complexity and variability of these systems.

    Existing research includes frameworks like DSPy, which optimizes LLM-based systems programmatically, and ProTeGi, which uses textual gradients for prompt optimization. DSPy enhances LLM performance in various tasks by structuring complex systems as layered programs. ProTeGi focuses on improving prompts through natural language feedback. These methods automate the optimization process but are limited to specific applications. TEXTGRAD, inspired by these approaches, expands the use of textual gradients to broader optimization tasks, integrating LLMs’ reasoning capabilities across diverse domains.

    Researchers from Stanford University and the Chan Zuckerberg Biohub have introduced TEXTGRAD, a framework that performs automatic differentiation via text, using feedback from LLMs to optimize AI systems. TEXTGRAD converts each AI system into a computation graph, where variables are inputs and outputs of complex functions. It leverages the rich, interpretable natural language feedback provided by LLMs to generate “textual gradients,” which describe how variables should be adjusted to improve system performance. This innovative approach makes TEXTGRAD flexible and easy to use, as users only need to provide the objective function without tuning components or prompts.

    TEXTGRAD employs LLMs to generate detailed feedback for various tasks, making the framework applicable across multiple domains. For instance, in the field of coding, TEXTGRAD improved the performance of AI models on difficult coding problems from the LeetCode platform. By identifying edge cases that caused failures in initial solutions, TEXTGRAD provided suggestions for improvement, leading to a 20% relative performance gain. In question-answering tasks, TEXTGRAD enhanced the zero-shot accuracy of GPT-4 in the Google-Proof Question Answering benchmark from 51% to 55%. The framework also designed new drug-like molecules with desirable properties, significantly improving binding affinity and drug-likeness metrics.

    TEXTGRAD’s results speak for themselves. In coding optimization, it improved the success rate of GPT-4 from 7% to 23% in a zero-shot setting and from 15% to 31% when using Reflexion. In problem-solving tasks, it boosted the accuracy of GPT-4 in the Google-Proof Question Answering benchmark to 55%, the highest known result for this dataset. For the Multi-Task Language Understanding (MMLU) benchmark, it increased the accuracy from 85.7% to 88.4% in the Machine Learning subset and from 91.2% to 95.1% in the College Physics subset. These impressive results underscore the effectiveness of TEXTGRAD in improving AI performance.

    TEXTGRAD optimized molecules for better binding affinity and drug-likeness in chemistry, demonstrating its versatility in multi-objective optimization tasks. The framework generated molecules with high binding affinities and favorable drug-likeness scores comparable to clinically approved drugs. In medical applications, TEXTGRAD improved radiotherapy treatment plans by optimizing hyperparameters to target tumors better while minimizing damage to healthy tissues. The framework’s ability to provide meaningful guidance through textual gradients resulted in treatment plans that met clinical goals more effectively than traditional methods.

    In conclusion, TEXTGRAD represents a significant advancement in AI optimization, leveraging the capabilities of LLMs to provide detailed, natural language feedback. This approach enables efficient and effective optimization of complex AI systems, paving the way for developing next-generation AI technologies. Researchers from Stanford University and the Chan Zuckerberg Biohub have demonstrated that TEXTGRAD’s flexibility and ease of use make it a powerful tool for enhancing AI performance across various domains. By automating the optimization process, TEXTGRAD reduces the reliance on manual adjustments, accelerating the progress of AI research and applications.

    The post Researchers at Stanford Introduce TEXTGRAD: A Powerful AI Framework Performing Automatic “Differentiation” via Text appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleDeepening Safety Alignment in Large Language Models (LLMs)
    Next Article ShiftAddLLM: Accelerating Pretrained LLMs through Post-Training Shift-and-Add Reparameterization: Creating Efficient Multiplication-Free Models

    Related Posts

    Development

    February 2025 Baseline monthly digest

    May 18, 2025
    Artificial Intelligence

    Markus Buehler receives 2025 Washington Award

    May 18, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Microsoft’s support docs are urging Windows 10 users to get Windows 11

    Development

    6 books engineers should read

    Development

    Kong AI Gateway updated with features to reduce LLM hallucination and protect sensitive personal data

    Tech & Work

    I’ve enjoyed FBC: Firebreak in preview, and I’m looking forward to more sticky notes, leeches, and showers this summer

    News & Updates

    Highlights

    Artificial Intelligence

    FCA Just Dropped Big News on Live AI Testing for UK Firms

    April 30, 2025

    UK regulator to launch live AI testing lab to help firms go public. The Financial…

    French Business Email Database List

    January 20, 2025

    Vesta – Hosting Control Panel

    January 8, 2025

    This AI Paper from Tel Aviv University Introduces GASLITE: A Gradient-Based Method to Expose Vulnerabilities in Dense Embedding-Based Text Retrieval Systems

    January 7, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.