Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 16, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 16, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 16, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 16, 2025

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025

      Minecraft licensing robbed us of this controversial NFL schedule release video

      May 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      The power of generators

      May 16, 2025
      Recent

      The power of generators

      May 16, 2025

      Simplify Factory Associations with Laravel’s UseFactory Attribute

      May 16, 2025

      This Week in Laravel: React Native, PhpStorm Junie, and more

      May 16, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025
      Recent

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Meta AI Releases OpenEQA: The Open-Vocabulary Embodied Question Answering Benchmark

    Meta AI Releases OpenEQA: The Open-Vocabulary Embodied Question Answering Benchmark

    April 14, 2024

    Significant progress has been made in LLMs, or large-scale language models, which have absorbed a fundamental linguistic understanding of the environment. However, LLMs, despite their proficiency in historical knowledge and insightful responses, are severely deficient in real-time comprehension.

    Imagine a pair of trendy smart glasses or a home robot with an embedded AI agent as its brain. For such an agent to be effective, it must be able to interact with humans using simple, everyday language and utilize senses like vision to understand its surroundings. This is the ambitious goal that Meta AI is pursuing, presenting a significant research challenge.

    EQA, a method for testing an AI agent’s comprehension of its environment, has practical implications that extend beyond the realm of research. Even the most basic form of EQA can simplify everyday life. For instance, consider a scenario where you need to leave the house but can’t find your office badge. EQA could help you locate it. However, as Moravec’s paradox suggests, even the most advanced models of today still can’t match human performance in EQA. 

    As a pioneering effort, Meta has introduced the Open-Vocabulary Embodied Question Answering (OpenEQA) framework. This innovative metric is designed to assess an AI agent’s understanding of its environment through open-vocabulary inquiries, a novel approach in the field. The concept is akin to testing a person’s comprehension of a topic by asking them questions and analyzing their responses. 

    The first part of OpenEQA is episodic memory EQA, which requires an embodied AI agent to recall prior experiences to answer questions. The second part is active EQA, which requires the agent to actively seek out information from its surroundings to answer questions.

    This benchmark includes over 180 movies and scans of physical environments, and over 1,600 non-templated question-and-answer pairs provided by human annotators that reflect real-world scenarios. LLM-Match, an automated evaluation criteria for rating open vocabulary answers, is also included with OpenEQA. Blind user trials demonstrated that LLM-Match is as closely associated with humans as two people are with one another.

    The team found a significant gap between human performance (85.9%), even among the most effective models (GPT-4V at 48.5%), and OpenEQA’s benchmarking of various state-of-the-art vision+language foundation models (VLMs). Even the most advanced VLMs struggle with spatial understanding questions, suggesting that models that use visual information aren’t fully utilizing it. Instead, they rely on prior textual knowledge to answer visual questions. This indicates that embodied AI entities driven by these models still have a long way to go in perception and reasoning before they are ready for widespread use.

    OpenEQA integrates the capacity to respond in natural language with the ability to tackle difficult open-vocabulary queries. This produces an easy-to-understand metric showing environmental expertise while challenging foundational assumptions. Researchers hope academics can use OpenEQA, the first open-vocabulary benchmark for EQA, to monitor developments in scene interpretation and multimodal learning.

    Check out the Paper, Project, and Blog. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

    If you like our work, you will love our newsletter..

    Don’t Forget to join our 40k+ ML SubReddit

    The post Meta AI Releases OpenEQA: The Open-Vocabulary Embodied Question Answering Benchmark appeared first on MarkTechPost.

    Source: Read More 

    Hostinger
    Facebook Twitter Reddit Email Copy Link
    Previous ArticleAutonomous Domain-General Evaluation Models Enhance Digital Agent Performance: A Breakthrough in Adaptive AI Technologies
    Next Article Tableau vs Power BI: A Comparison of AI-Powered Analytics Tools

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 16, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-47916 – Invision Community Themeeditor Remote Code Execution

    May 16, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    CVE-2025-30386 – Microsoft Office Use After Free Code Execution Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    Microsoft Edge Joins Chrome in Matching Scrollbars to Your Theme

    Operating Systems

    Error’d: Killing Time

    News & Updates

    How to use AI coding tools to learn a new programming language

    Development

    Highlights

    Development

    Germany Attributes 2021 Attack On Federal Cartography Agency To China

    August 1, 2024

    The Federal Office for Cartography and Geodesy, known as the Bundesamt für Kartographie und Geodäsie…

    BreachForums Returns Just Weeks After FBI Seizure – Honeypot or Blunder?

    May 29, 2024

    Firefox 137 Released with Address Bar Revamp & Tab Groups

    April 1, 2025

    Why are GFCI outlets such a big deal? And 6 things you should never plug into one

    February 25, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.