Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 16, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 16, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 16, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 16, 2025

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025

      Minecraft licensing robbed us of this controversial NFL schedule release video

      May 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      The power of generators

      May 16, 2025
      Recent

      The power of generators

      May 16, 2025

      Simplify Factory Associations with Laravel’s UseFactory Attribute

      May 16, 2025

      This Week in Laravel: React Native, PhpStorm Junie, and more

      May 16, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025
      Recent

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Tsinghua University Researchers Propose ADELIE: Enhancing Information Extraction with Aligned Large Language Models Around Human-Centric Tasks

    Tsinghua University Researchers Propose ADELIE: Enhancing Information Extraction with Aligned Large Language Models Around Human-Centric Tasks

    May 12, 2024

    Information extraction (IE) is a pivotal area of artificial intelligence that transforms unstructured text into structured, actionable data. Despite their expansive capacities, traditional large language models (LLMs) often fail to comprehend and execute the nuanced directives required for precise IE. These challenges primarily manifest in closed IE tasks, where a model must adhere to stringent, pre-defined schemas.

    IE tasks compel models to discern and categorize text in formats that align with predefined structures, such as named entity recognition and relation classification. However, existing LLMs typically falter when tasked with the nuanced understanding and alignment necessary for effective IE. Researchers have traditionally employed strategies such as prompt engineering, which involves providing detailed annotations and guidelines to assist LLMs without altering underlying model parameters.

    The research community has observed a critical need for a methodology that enhances LLMs’ understanding of structured tasks and improves execution accuracy. In response, researchers from Tsinghua University have introduced a new approach called ADELIE (Aligning large language moDELs on Information Extraction). This approach leverages a specialized dataset, IEInstruct, comprising over 83,000 instances across various IE formats, including triplets, natural language responses, and JSON outputs. 

    ADELIE diverges from conventional methods by integrating supervised fine-tuning with an innovative Direct Preference Optimization (DPO) strategy. This blend enables the model to align more closely with the intricacies of human-like IE processing. Initial training involves a mix of IE-specific and generic data, using the LLAMA 2 model over 6,306 gradient steps, which ensures the retention of broad linguistic capabilities alongside specialized IE performance.

    Performance metrics reveal that ADELIE models, ADELIESFT and ADELIEDPO, achieve benchmark-setting results. In evaluations against held-out datasets, ADELIESFT shows an average F1 score improvement of 5% over standard LLM outputs in closed IE tasks. The improvements are even more pronounced for open IE, with ADELIE models outperforming state-of-the-art alternatives by 3-4% margins in robustness and extraction accuracy. In the realm of on-demand IE, the models demonstrate a nuanced understanding of user instructions, translating into highly accurate data structuring.

    In conclusion, ADELIE’s methodical training and optimization translate into a potent alignment of LLMs with IE tasks, demonstrating that a focused approach to data diversity and instruction specificity can bridge the gap between human expectations and machine performance. This alignment does not compromise the models’ general capabilities, which is often a concern with task-specific tuning. The impressive results across various metrics and task types underscore the potential of ADELIE to set new standards in information extraction, making it a valuable tool for multiple applications, from academic research to real-world data processing.

    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

    If you like our work, you will love our newsletter..

    Don’t Forget to join our 42k+ ML SubReddit

    The post Tsinghua University Researchers Propose ADELIE: Enhancing Information Extraction with Aligned Large Language Models Around Human-Centric Tasks appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleThis AI Paper by the University of Michigan Introduces MIDGARD: Advancing AI Reasoning with Minimum Description Length
    Next Article UC Berkeley Researchers Introduce Learnable Latent Codes as Bridges (LCB): A Novel AI Approach that Combines the Abstract Reasoning Capabilities of Large Language Models with Low-Level Action Policies

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 17, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-48187 – RAGFlow Authentication Bypass

    May 17, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    The Secret of the World’s Darkest Prison

    Artificial Intelligence

    It’s Time to Disrupt Visual Regression Testing

    Development

    Smashing Animations Part 1: How Classic Cartoons Inspire Modern CSS

    Tech & Work

    Essential Tools and Frameworks for Mastering Ethical Hacking on Linux

    Learning Resources

    Highlights

    Black Hat Europe 2024: Hacking a car – or rather, its infotainment system

    December 20, 2024

    Our ‘computers on wheels’ are more connected than ever, but the features that enhance our…

    Edward Snowden labels OpenAI’s new board appointment a “willful, calculated betrayal of the rights of every person on Earth”

    June 24, 2024

    Rilasciata Finnix 250: La distribuzione GNU/Linux per amministratori di sistema celebra 25 anni

    March 23, 2025

    Middle East’s Top 100 Cybersecurity Leaders to Follow

    August 13, 2024
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.