Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 31, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 31, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 31, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 31, 2025

      Windows 11 version 25H2: Everything you need to know about Microsoft’s next OS release

      May 31, 2025

      Elden Ring Nightreign already has a duos Seamless Co-op mod from the creator of the beloved original, and it’ll be “expanded on in the future”

      May 31, 2025

      I love Elden Ring Nightreign’s weirdest boss — he bargains with you, heals you, and throws tantrums if you ruin his meditation

      May 31, 2025

      How to install SteamOS on ROG Ally and Legion Go Windows gaming handhelds

      May 31, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Oracle Fusion new Product Management Landing Page and AI (25B)

      May 31, 2025
      Recent

      Oracle Fusion new Product Management Landing Page and AI (25B)

      May 31, 2025

      Filament Is Now Running Natively on Mobile

      May 31, 2025

      How Remix is shaking things up

      May 30, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Windows 11 version 25H2: Everything you need to know about Microsoft’s next OS release

      May 31, 2025
      Recent

      Windows 11 version 25H2: Everything you need to know about Microsoft’s next OS release

      May 31, 2025

      Elden Ring Nightreign already has a duos Seamless Co-op mod from the creator of the beloved original, and it’ll be “expanded on in the future”

      May 31, 2025

      I love Elden Ring Nightreign’s weirdest boss — he bargains with you, heals you, and throws tantrums if you ruin his meditation

      May 31, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»Meta AI Introduces CLUE (Constitutional MLLM JUdgE): An AI Framework Designed to Address the Shortcomings of Traditional Image Safety Systems

    Meta AI Introduces CLUE (Constitutional MLLM JUdgE): An AI Framework Designed to Address the Shortcomings of Traditional Image Safety Systems

    January 13, 2025

    The rapid growth of digital platforms has brought image safety into sharp focus. Harmful imagery—ranging from explicit content to depictions of violence—poses significant challenges for content moderation. The proliferation of AI-generated content (AIGC) has exacerbated these challenges, as advanced image-generation models can easily create unsafe visuals. Current safety systems rely heavily on human-labeled datasets, which are both expensive and difficult to scale. Moreover, these systems often struggle to adapt to evolving and complex safety guidelines. An effective solution must address these limitations while ensuring efficient and reliable image safety assessments.

    Researchers from Meta, Rutgers University, Westlake University, and UMass Amherst have developed CLUE (Constitutional MLLM JUdgE), a framework designed to address the shortcomings of traditional image safety systems. CLUE uses Multimodal Large Language Models (MLLMs) to convert subjective safety rules into objective, measurable criteria. Key features of the framework include:

    1. Constitution Objectification: Converting subjective safety rules into clear, actionable guidelines for better processing by MLLMs.
    2. Rule-Image Relevance Checks: Leveraging CLIP to efficiently filter irrelevant rules by assessing the relevance between images and guidelines.
    3. Precondition Extraction: Breaking down complex rules into simplified precondition chains for easier reasoning.
    4. Debiased Token Probability Analysis: Mitigating biases caused by language priors and non-central image regions to improve objectivity.
    5. Cascaded Reasoning: Employing deeper chain-of-thought reasoning for cases with low confidence to enhance decision-making accuracy.

    Technical Details and Benefits

    The CLUE framework addresses key challenges associated with MLLMs in image safety. By objectifying safety rules, it replaces ambiguous guidelines with precise criteria, such as specifying “should not depict people with visible, bloody injuries indicating imminent death.”

    Relevance scanning using CLIP streamlines the process by removing rules irrelevant to the inspected image, thus reducing computational load. This ensures the framework focuses only on pertinent rules, improving efficiency.

    The precondition extraction module simplifies complex rules into logical components, enabling MLLMs to reason more effectively. For example, a rule like “should not depict any people whose bodies are on fire” is decomposed into conditions such as “people are visible” and “bodies are on fire.”

    Debiased token probability analysis is another notable feature. By comparing token probabilities with and without image tokens, biases are identified and minimized. This reduces the likelihood of errors, such as associating background elements with violations.

    The cascaded reasoning mechanism provides a robust fallback for low-confidence scenarios. Using step-by-step logical reasoning, it ensures accurate assessments, even for borderline cases, while offering detailed justifications for decisions.

    Experimental Results and Insights

    CLUE’s effectiveness has been validated through extensive testing on various MLLM architectures, including InternVL2-76B, Qwen2-VL-7B-Instruct, and LLaVA-v1.6-34B. Key findings include:

    • Accuracy and Recall: CLUE achieved 95.9% recall and 94.8% accuracy with InternVL2-76B, outperforming existing methods.
    • Efficiency: The relevance scanning module filtered out 67% of irrelevant rules while retaining 96.6% of ground-truth violated rules, significantly improving computational efficiency.
    • Generalizability: Unlike fine-tuned models, CLUE performed well across diverse safety guidelines, highlighting its scalability.

    Insights also emphasize the importance of constitution objectification and debiased token probability analysis. Objectified rules achieved a 98.0% accuracy rate compared to 74.0% for their original counterparts, underlining the value of clear and measurable criteria. Similarly, debiasing improved overall judgment accuracy, with an F1-score of 0.879 for the InternVL2-8B-AWQ model.

    Conclusion

    CLUE offers a thoughtful and efficient approach to image safety, addressing the limitations of traditional methods by leveraging MLLMs. By transforming subjective rules into objective criteria, filtering irrelevant rules, and utilizing advanced reasoning mechanisms, CLUE provides reliable and scalable solutions for content moderation. Its ability to deliver high accuracy and adaptability makes it a significant advancement in managing the challenges of AI-generated content, paving the way for safer online platforms.


    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 65k+ ML SubReddit.

    🚨 FREE UPCOMING AI WEBINAR (JAN 15, 2025): Boost LLM Accuracy with Synthetic Data and Evaluation Intelligence–Join this webinar to gain actionable insights into boosting LLM model performance and accuracy while safeguarding data privacy.

    The post Meta AI Introduces CLUE (Constitutional MLLM JUdgE): An AI Framework Designed to Address the Shortcomings of Traditional Image Safety Systems appeared first on MarkTechPost.

    Source: Read More 

    Hostinger
    Facebook Twitter Reddit Email Copy Link
    Previous ArticleSalesforce AI Introduces TACO: A New Family of Multimodal Action Models that Combine Reasoning with Real-World Actions to Solve Complex Visual Tasks
    Next Article Researchers from Fudan University and Shanghai AI Lab Introduces DOLPHIN: A Closed-Loop Framework for Automating Scientific Research with Iterative Feedback

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    May 31, 2025
    Machine Learning

    Cisco’s Latest AI Agents Report Details the Transformative Impact of Agentic AI on Customer Experience

    May 31, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Enhancing UI Design with the Gestalt Principle of Common Fate

    Development

    CVE-2025-3514 – “SureForms WordPress Stored Cross-Site Scripting Vulnerability”

    Common Vulnerabilities and Exposures (CVEs)

    Freeimage.dll – Do You Need It? How to Remove

    Development

    Wearable Accelerometer Foundation Models for Health via Knowledge Distillation

    Machine Learning

    Highlights

    Development

    DeepSeek AI Researchers Propose Expert-Specialized Fine-Tuning, or ESFT to Reduce Memory by up to 90% and Time by up to 30%

    July 6, 2024

    Natural language processing is advancing rapidly, focusing on optimizing large language models (LLMs) for specific…

    Meet OpenShift Lightspeed, RedHat’s AI tool for Kubernetes admins

    August 6, 2024

    FACTS Grounding: A new benchmark for evaluating the factuality of large language models

    December 20, 2024

    CVE-2025-44862 – TOTOLINK CA300-POE Command Injection Vulnerability

    May 1, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.