Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      June 4, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      June 4, 2025

      How To Prevent WordPress SQL Injection Attacks

      June 4, 2025

      Smashing Animations Part 4: Optimising SVGs

      June 4, 2025

      I test AI tools for a living. Here are 3 image generators I actually use and how

      June 4, 2025

      The world’s smallest 65W USB-C charger is my latest travel essential

      June 4, 2025

      This Spotlight alternative for Mac is my secret weapon for AI-powered search

      June 4, 2025

      Tech prophet Mary Meeker just dropped a massive report on AI trends – here’s your TL;DR

      June 4, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Beyond AEM: How Adobe Sensei Powers the Full Enterprise Experience

      June 4, 2025
      Recent

      Beyond AEM: How Adobe Sensei Powers the Full Enterprise Experience

      June 4, 2025

      Simplify Negative Relation Queries with Laravel’s whereDoesntHaveRelation Methods

      June 4, 2025

      Cast Model Properties to a Uri Instance in 12.17

      June 4, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      My Favorite Obsidian Plugins and Their Hidden Settings

      June 4, 2025
      Recent

      My Favorite Obsidian Plugins and Their Hidden Settings

      June 4, 2025

      Rilasciata /e/OS 3.0: Nuova Vita per Android Senza Google, Più Privacy e Controllo per l’Utente

      June 4, 2025

      Rilasciata Oracle Linux 9.6: Scopri le Novità e i Miglioramenti nella Sicurezza e nelle Prestazioni

      June 4, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»ARM: Enhancing Open-Domain Question Answering with Structured Retrieval and Efficient Data Alignment

    ARM: Enhancing Open-Domain Question Answering with Structured Retrieval and Efficient Data Alignment

    February 3, 2025

    Answering open-domain questions in real-world scenarios is challenging, as relevant information is often scattered across diverse sources, including text, databases, and images. While LLMs can break down complex queries into simpler steps to improve retrieval, they usually fail to account for how data is structured, leading to suboptimal results. Agentic RAG introduces iterative retrieval, refining searches based on prior results. However, this approach is inefficient, as queries are guided by past retrievals rather than data organization. Additionally, it lacks joint optimization, making it prone to reasoning derailment, where errors in early steps cascade into incorrect decisions, increasing computational costs.

    Researchers from MIT, AWS AI, and the University of Pennsylvania introduced ARM, an LLM-based retrieval method designed to enhance complex question answering by aligning queries with the structure of available data. Unlike conventional approaches, ARM explores relationships between data objects rather than relying solely on semantic matching, enabling a retrieve-all-at-once solution. Evaluated on Bird and OTT-QA datasets, ARM outperformed standard RAG and agentic RAG, achieving up to 5.2 and 15.9 points higher execution accuracy on Bird and up to 5.5 and 19.3 points higher F1 scores on OTT-QA. ARM improves retrieval efficiency through structured reasoning and alignment verification.

    The alignment-driven LLM retrieval framework integrates retrieval and reasoning within a unified decoding process, optimizing it through beam search. Unlike conventional methods that treat retrieval and reasoning as separate steps, the LLM can dynamically retrieve relevant data objects while incorporating structured data, a reasoning solver, and self-verification. Since LLMs lack direct access to structured data, we frame retrieval as a generative task, where the model formulates reasoning to identify essential data objects. This process involves iterative decoding with three key components: information alignment, structure alignment, and self-verification, ensuring logical consistency and accurate retrieval.

    Textual data is indexed as N-grams and embeddings to enhance retrieval accuracy, enabling constrained beam decoding for precise alignment. Information alignment extracts key terms and retrieves relevant objects using BM25 scoring and embedding-based similarity. Structure alignment refines these objects through an optimization model, ensuring logical coherence. Finally, self-verification allows the LLM to validate and integrate selected objects within a structured reasoning framework. Multiple drafts are generated through controlled object expansion, and beam search aggregation prioritizes the most confident selections, ensuring high-quality, contextually relevant responses from diverse data sources.

    The study assesses the method on open-domain question-answering tasks using OTT-QA and Bird datasets. OTT-QA involves short-text answers from passages and tables, while Bird requires SQL queries from multiple tables. We compare our approach with standard and agentic RAG baselines, incorporating query decomposition and reranking. ARM, using Llama-3.1-8B-Instruct, retrieves relevant objects efficiently, outperforming baselines in recall and end-to-end accuracy while reducing LLM calls. ReAct struggles with iterative reasoning errors, often repeating searches. ARM’s structured retrieval process improves precision and efficiency. The results highlight ARM’s superiority in retrieving essential information while maintaining computational efficiency across both datasets.

    In conclusion, Effective open-domain question answering requires understanding the available data objects and their organization. Query decomposition with an off-the-shelf LLM often leads to suboptimal retrieval due to a lack of awareness about the data structure. While agentic RAG can interact with the data, it relies on previous retrieval results, making it inefficient and increasing LLM calls. The proposed ARM retrieval method identifies and navigates relevant data objects, even those not directly mentioned in the question. Experimental results show that ARM outperforms baselines in retrieval accuracy and efficiency, requiring fewer LLM calls for improved performance in downstream tasks.


    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 75k+ ML SubReddit.

    🚨 Marktechpost is inviting AI Companies/Startups/Groups to partner for its upcoming AI Magazines on ‘Open Source AI in Production’ and ‘Agentic AI’.

    The post ARM: Enhancing Open-Domain Question Answering with Structured Retrieval and Efficient Data Alignment appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleThis AI Paper from Meta Introduces Diverse Preference Optimization (DivPO): A Novel Optimization Method for Enhancing Diversity in Large Language Models
    Next Article Accelerate video Q&A workflows using Amazon Bedrock Knowledge Bases, Amazon Transcribe, and thoughtful UX design

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    June 4, 2025
    Machine Learning

    A Coding Implementation to Build an Advanced Web Intelligence Agent with Tavily and Gemini AI

    June 4, 2025
    Leave A Reply Cancel Reply

    Hostinger

    Continue Reading

    Warhammer 40,000: Rogue Trader, winner of Windows Central’s 2024 Editor’s Choice Award, has sold over 1 million copies

    News & Updates

    Apple to Introduce AI-Powered Safari Browser Alongside New Operating Systems

    Artificial Intelligence

    ChatGPT Can Now Schedule Tasks in Advance

    News & Updates

    10 Best WordPress Plugins to Try Out in 2025

    News & Updates
    Hostinger

    Highlights

    Linux Candy: Meme is a fun tool to create memes

    April 10, 2025

    Meme is a command line utility for creating image macro style memes. It’s free and…

    Google AI Introduce AGREE: A Machine Learning Framework that Enables LLMs to Self-Ground the Claims in their Responses and to Provide Precise Citations

    May 29, 2024

    With 700,000 Large Language Models (LLMs) On Hugging Face Already, Where Is The Future of Artificial Intelligence AI Headed?

    June 15, 2024

    NVIDIA AI Introduces Omni-RGPT: A Unified Multimodal Large Language Model for Seamless Region-level Understanding in Images and Videos

    January 17, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.