Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 16, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 16, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 16, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 16, 2025

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025

      Minecraft licensing robbed us of this controversial NFL schedule release video

      May 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      The power of generators

      May 16, 2025
      Recent

      The power of generators

      May 16, 2025

      Simplify Factory Associations with Laravel’s UseFactory Attribute

      May 16, 2025

      This Week in Laravel: React Native, PhpStorm Junie, and more

      May 16, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025
      Recent

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»This AI Paper from UCSD and CMU Introduces EDU-RELAT: A Benchmark for Evaluating Deep Unlearning in Large Language Models

    This AI Paper from UCSD and CMU Introduces EDU-RELAT: A Benchmark for Evaluating Deep Unlearning in Large Language Models

    December 8, 2024

    Large language models (LLMs) excel in generating contextually relevant text; however, ensuring compliance with data privacy regulations, such as GDPR, requires a robust ability to unlearn specific information effectively. This capability is critical for addressing privacy concerns where data must be entirely removed from models and any logical connections that could reconstruct deleted information.

    The problem of unlearning in LLMs is particularly challenging due to the interconnected nature of knowledge stored within these models. Removing a single fact is insufficient if related or deduced facts remain. For instance, removing a family relationship fact does not prevent the model from inferring it through logical rules or remaining connections. Addressing this issue necessitates unlearning approaches that consider explicit data and its logical dependencies.

    Current unlearning methods focus on removing specific data points, such as gradient ascent, negative preference optimization (NPO), and task vector methods. These approaches aim to erase data while retaining overall model utility. However, they need to achieve deep unlearning, which involves removing the target fact and any inferable connections. This limitation compromises the completeness of data erasure and may lead to collateral damage, erasing unrelated facts unnecessarily.

    Researchers from the University of California, San Diego, & Carnegie Mellon University introduced the concept of “deep unlearning” to address these challenges. The research leverages a synthetic dataset, EDU-RELAT, consisting of 400 family relationships, 300 biographical facts, and 48 logical rules. These serve as a benchmark for evaluating unlearning methods. Key metrics such as recall, which measures the extent of unlearning, and accuracy, which assesses the preservation of unrelated facts, were used to evaluate performance.

    In the study, the researchers tested four unlearning techniques—Gradient Ascent (GA), Negative Preference Optimization (NPO), Task Vector (TV), and Who’s Harry Potter (WHP)—on four prominent LLMs: GPT2-XL, Phi-1.5, Llama2-7b, and Llama3-8b. The evaluation focused on deeply unlearning 55 specific facts related to family relationships. The study measured both Acc@Recall ≥ 0.8 (accuracy when a recall is above 80%) and Recall@Acc ≥ 0.8 (recall when accuracy is above 80%) to balance comprehensiveness and utility.

    The results highlighted significant areas for improvement in existing unlearning methods—none of the methods achieved high recall and accuracy. For example, Gradient Ascent achieved a recall of 75% on Llama2-7b but often caused collateral damage by unlearning unrelated facts. NPO and Task Vector achieved recall rates between 70%-73% on larger models, such as Llama3-8b. In contrast, WHP performed poorly, with recall rates below 50% across all models. Larger models like Llama2-7b and Llama3-8b outperformed smaller ones like GPT2-XL and Phi-1.5 due to their more advanced reasoning capabilities, which aided in better handling logical dependencies.

    Additional analysis revealed that the accuracy of biographical facts was generally higher than that of family relationships. For instance, GA achieved an Acc@Recall ≥ 0.8 for biographical facts on Llama2-7b and Llama3-8b but only 0.6 for family relationships. This discrepancy highlights the difficulty in unlearning closely related facts without unintended losses. Unlearning a single fact often required removing ten or more related facts, demonstrating the complexity of deep unlearning.

    The research underscores the limitations of current approaches in achieving effective deep unlearning. While methods like Gradient Ascent showed potential for superficial unlearning, they needed to be more for deeply interconnected datasets. The study concludes with a call for comprehensively developing new methodologies that address these challenges. By establishing deep unlearning as a benchmark, this research lays the groundwork for advancements in privacy-preserving AI and highlights the importance of balancing unlearning efficacy with model utility.


    Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 60k+ ML SubReddit.

    🚨 [Must Attend Webinar]: ‘Transform proofs-of-concept into production-ready AI applications and agents’ (Promoted)

    The post This AI Paper from UCSD and CMU Introduces EDU-RELAT: A Benchmark for Evaluating Deep Unlearning in Large Language Models appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleMicrosoft Introduces Florence-VL: A Multimodal Model Redefining Vision-Language Alignment with Generative Vision Encoding and Depth-Breadth Fusion
    Next Article Oregon Ducks Blue 84 2024 Big Ten Football Conference Champions Shirt

    Related Posts

    Machine Learning

    LLMs Struggle with Real Conversations: Microsoft and Salesforce Researchers Reveal a 39% Performance Drop in Multi-Turn Underspecified Tasks

    May 17, 2025
    Machine Learning

    This AI paper from DeepSeek-AI Explores How DeepSeek-V3 Delivers High-Performance Language Modeling by Minimizing Hardware Overhead and Maximizing Computational Efficiency

    May 17, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Rilasciata Chimera Linux 20250214: Un Passo Avanti Verso la Stabilità

    Linux

    Casino Online Argentina Mercadopago Mejores Casinos En Mercadopago

    Development

    CVE-2018-1359 – Apache HTTP Server Authentication Bypass

    Common Vulnerabilities and Exposures (CVEs)

    The 20+ best early Amazon Prime Day deals under $50

    Development

    Highlights

    AI Desk offers 24/7 AI support to cut costs

    December 27, 2024

    Post Content Source: Read More 

    AMD’s new Radeon graphics driver is ready for DOOM: The Dark Ages and delivers FSR 4 upscaling support to Frostpunk 2, more

    May 10, 2025

    Apple Intelligence page no longer reads “Available Now” — I wonder if Apple has lost faith in its AI strategy after so many delays?

    April 24, 2025

    The Dumbest Thing in Security This Week: CrowdStrike, Delta and Information Asymmetry

    August 3, 2024
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.