Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 18, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 18, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 18, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 18, 2025

      I need to see more from Lenovo’s most affordable gaming desktop, because this isn’t good enough

      May 18, 2025

      Gears of War: Reloaded — Release date, price, and everything you need to know

      May 18, 2025

      I’ve been using the Logitech MX Master 3S’ gaming-influenced alternative, and it could be your next mouse

      May 18, 2025

      Your Android devices are getting several upgrades for free – including a big one for Auto

      May 18, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      YTConverter™ lets you download YouTube videos/audio cleanly via terminal — especially great for Termux users.

      May 18, 2025
      Recent

      YTConverter™ lets you download YouTube videos/audio cleanly via terminal — especially great for Termux users.

      May 18, 2025

      NodeSource N|Solid Runtime Release – May 2025: Performance, Stability & the Final Update for v18

      May 17, 2025

      Big Changes at Meteor Software: Our Next Chapter

      May 17, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      I need to see more from Lenovo’s most affordable gaming desktop, because this isn’t good enough

      May 18, 2025
      Recent

      I need to see more from Lenovo’s most affordable gaming desktop, because this isn’t good enough

      May 18, 2025

      Gears of War: Reloaded — Release date, price, and everything you need to know

      May 18, 2025

      I’ve been using the Logitech MX Master 3S’ gaming-influenced alternative, and it could be your next mouse

      May 18, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»TransFusion: An Artificial Intelligence AI Framework To Boost a Large Language Model’s Multilingual Instruction-Following Information Extraction Capability

    TransFusion: An Artificial Intelligence AI Framework To Boost a Large Language Model’s Multilingual Instruction-Following Information Extraction Capability

    June 30, 2024

    Large Language Models (LLMs) have made significant advances in the field of Information Extraction (IE). Information extraction is a task in Natural Language Processing (NLP) that involves identifying and extracting specific pieces of information from text. LLMs have demonstrated great results in IE, especially when combined with instruction tuning. Through instruction tuning, these models are trained to annotate text according to predetermined standards, which improves their ability to generalize to new datasets. This indicates that even with unknown data, people are able to do IE tasks successfully by following instructions.

    However, even with these improvements, LLMs still face many difficulties when working with low-resource languages. These languages lack both the unlabeled text required for pre-training and the labeled data required for fine-tuning models. Due to this lack of data, it is challenging for LLMs to attain good performance in these languages.

    To overcome this, a team of researchers from the Georgia Institute of Technology has introduced the TransFusion framework. In TransFusion, models are adjusted to function with data translated from low-resource languages into English. With this method, the original low-resource language text and its English translation provide information that the models may use to create more accurate predictions.

    This framework aims to effectively enhance IE in low-resource languages by utilizing external Machine Translation (MT) systems. There are three primary steps involved, which are as follows:

    Translation during Inference: Converting low-resource language data into English so that a high-resource model can annotate it.

    Fusion of Annotated Data: In a model trained to use both types of data, fusing the original low-resource language text with the annotated English translations.

    Constructing a TransFusion Reasoning Chain, which integrates both annotation and fusion into a single autoregressive decoding pass.

    Expanding upon this structure, the team has also introduced GoLLIE-TF, which is an instruction-tuned LLM that is cross-lingual and tailored especially for Internet Explorer tasks. GoLLIE-TF aims to reduce the performance disparity between high- and low-resource languages. The combined goal of the TransFusion framework and GoLLIE-TF is to increase LLMs’ efficiency when handling low-resource languages.

    Experiments on twelve multilingual IE datasets, with a total of fifty languages, have shown that GoLLIE-TF works well. In comparison to the basic model, the results demonstrate that GoLLIE-TF performs greater zero-shot cross-lingual transfer. This means that without further training data, it can more effectively apply its acquired skills to new languages.

    TransFusion applied to proprietary models such as GPT-4 considerably improves the performance of low-resource language named entity recognition (NER). When prompting was used, GPT-4’s performance increased by 5 F1 points. Further improvements were obtained by fine-tuning various language model types using the TransFusion framework; decoder-only architectures improved by 14 F1 points, while encoder-only designs improved by 13 F1 points.

    In conclusion, TransFusion and GoLLIE-TF together provide a potent solution for enhancing IE tasks in low-resource languages. This shows notable improvements across many models and datasets, helping to reduce the performance gap between high-resource and low-resource languages by utilizing English translations and fine-tuning models to fuse annotations.

    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. 

    Join our Telegram Channel and LinkedIn Group.

    If you like our work, you will love our newsletter..

    Don’t Forget to join our 45k+ ML SubReddit

    The post TransFusion: An Artificial Intelligence AI Framework To Boost a Large Language Model’s Multilingual Instruction-Following Information Extraction Capability appeared first on MarkTechPost.

    Source: Read More 

    Hostinger
    Facebook Twitter Reddit Email Copy Link
    Previous Article10 Use Cases of Claude 3.5 Sonnet: Unveiling the Future of Artificial Intelligence AI with Revolutionary Capabilities
    Next Article Llama-Agents: A New Open-Source AI Framework that Simplifies the Creation, Iteration, and Deployment of Multi-Agent AI Systems

    Related Posts

    Development

    February 2025 Baseline monthly digest

    May 18, 2025
    Artificial Intelligence

    Markus Buehler receives 2025 Washington Award

    May 18, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    How BlackSuit Ransomware is Crippling Businesses: FBI, CISA Sound Alarm

    Development

    Top 10 Highly Paid Indian CEOs in the USA

    Development

    CVE-2025-3852 – WordPress WPshop E-Commerce Privilege Escalation Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    ASUS Urges Users to Patch AiCloud Router Vuln Immediately

    Security

    Highlights

    Your push was rejected because it contains files larger than 10 MiB

    January 24, 2025

    I am working on a AI model that have a few files larger than 10MiB…

    Collective #892

    December 27, 2024

    New to open source? Here’s everything you need to get started

    January 24, 2025

    CVE-2024-56523 – Radware Cloud Web Application Firewall (WAF) HTTP Request Smuggling

    May 12, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.