Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 18, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 18, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 18, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 18, 2025

      I need to see more from Lenovo’s most affordable gaming desktop, because this isn’t good enough

      May 18, 2025

      Gears of War: Reloaded — Release date, price, and everything you need to know

      May 18, 2025

      I’ve been using the Logitech MX Master 3S’ gaming-influenced alternative, and it could be your next mouse

      May 18, 2025

      Your Android devices are getting several upgrades for free – including a big one for Auto

      May 18, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      YTConverter™ lets you download YouTube videos/audio cleanly via terminal — especially great for Termux users.

      May 18, 2025
      Recent

      YTConverter™ lets you download YouTube videos/audio cleanly via terminal — especially great for Termux users.

      May 18, 2025

      NodeSource N|Solid Runtime Release – May 2025: Performance, Stability & the Final Update for v18

      May 17, 2025

      Big Changes at Meteor Software: Our Next Chapter

      May 17, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      I need to see more from Lenovo’s most affordable gaming desktop, because this isn’t good enough

      May 18, 2025
      Recent

      I need to see more from Lenovo’s most affordable gaming desktop, because this isn’t good enough

      May 18, 2025

      Gears of War: Reloaded — Release date, price, and everything you need to know

      May 18, 2025

      I’ve been using the Logitech MX Master 3S’ gaming-influenced alternative, and it could be your next mouse

      May 18, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»This AI Paper Presents SliCK: A Knowledge Categorization Framework for Mitigating Hallucinations in Language Models Through Structured Training

    This AI Paper Presents SliCK: A Knowledge Categorization Framework for Mitigating Hallucinations in Language Models Through Structured Training

    May 14, 2024

    Research in computational linguistics continues to explore how large language models (LLMs) can be adapted to integrate new knowledge without compromising the integrity of existing information. A key challenge is ensuring that these models, fundamental to various language processing applications, maintain accuracy even as they expand their knowledge bases. 

    One conventional approach involves supervised fine-tuning, where LLMs are incrementally trained on data that aligns with or extends beyond their pre-training. While popular, this method has shown mixed results. The fine-tuning process involves presenting the model with examples it might partially recognize or not know, prompting it to adjust its responses accordingly. The effectiveness of these methods is typically evaluated by how these models maintain their performance when presented with data that either aligns with or extends their existing knowledge base.

    A research team from Technion – Israel Institute of Technology and Google Research has introduced SliCK, a novel framework specifically designed to examine integrating new knowledge within LLMs. This methodology stands out by categorizing knowledge into distinct levels, ranging from HighlyKnown to Unknown, providing a granular analysis of how different types of information affect model performance. This setup allows for a precise evaluation of the model’s ability to assimilate new facts while maintaining the accuracy of its existing knowledge base, highlighting the delicate balance required in model training.

    In the methodology, the study leverages the PaLM model, a robust LLM developed by Google, which was fine-tuned using datasets carefully designed to include varying proportions of knowledge categories: HighlyKnown, MaybeKnown, WeaklyKnown, and Unknown. These datasets are derived from a curated subset of factual questions mapped from Wikidata relations, enabling a controlled examination of the model’s learning dynamics. The experiment meticulously quantifies the model’s performance across these categories using exact match (EM) metrics to assess how effectively the model integrates new information while avoiding the pitfalls of hallucinations. This structured approach provides a clear view of the impact of fine-tuning with both familiar and novel data on model accuracy.

    The study’s findings demonstrate the effectiveness of the SliCK categorization in enhancing the fine-tuning process. Models trained using this structured approach, particularly with a 50% Known and 50% Unknown mix, showed an optimized balance, achieving a 5% higher accuracy in generating correct responses compared to models trained with predominantly Unknown data. Conversely, when the proportion of Unknown data exceeded 70%, the models’ propensity for hallucinations increased by approximately 12%. These results highlight SliCK’s critical role in quantitatively assessing and managing the risk of error as new information is integrated during the fine-tuning of LLMs.

    To summarize, the research by Technion – Israel Institute of Technology and Google Research thoroughly examines fine-tuning LLMs using the SliCK framework to manage the integration of new knowledge. The study highlights the delicate balance required in model training, with the PaLM model demonstrating improved accuracy and reduced hallucinations when trained under controlled knowledge conditions. These findings underscore the importance of strategic data categorization in enhancing model reliability and performance, offering valuable insights for future developments in machine learning methodologies.

    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

    If you like our work, you will love our newsletter..

    Don’t Forget to join our 42k+ ML SubReddit

    The post This AI Paper Presents SliCK: A Knowledge Categorization Framework for Mitigating Hallucinations in Language Models Through Structured Training appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleRadOnc-GPT: Leveraging Meta Llama for a Pioneering Radiation Oncology Model
    Next Article Generative AI in Marketing and Sales: A Comprehensive Review

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 18, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-4873 – PHPGurukul News Portal SQL Injection Vulnerability

    May 18, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Apple’s 18.8-inch foldable might hit the market in 2027 and run macOS

    Operating Systems

    Node 22.5.0 now includes node:sqlite module (22.5.1 bugfix)

    Development

    CVE-2024-56006 – Automattic Jetpack Debug Tools Missing Authorization Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    Part 1: Capture and Monitor Critical Power Automate Workflows History through Child Flow

    Development

    Highlights

    OpenAI decides to reverse recent GPT-4o update after user find bot being overly appeasing

    April 28, 2025

    OpenAI’s CEO, Sam Altman, announced the company will reverse the latest GPT-4o update after users…

    Chris Pashley Joins ARPA-H as Chief Information Security Officer

    June 20, 2024

    Global-MMLU: A World-class Benchmark Redefining Multilingual AI by Bridging Cultural and Linguistic Gaps for Equitable Evaluation Across 42 Languages and Diverse Contexts

    December 7, 2024

    India’s Average Data Breach Cost $2.18 Million in 2023: RBI Cybersecurity Report

    July 29, 2024
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.