Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 13, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 13, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 13, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 13, 2025

      This $4 Steam Deck game includes the most-played classics from my childhood — and it will save you paper

      May 13, 2025

      Microsoft shares rare look at radical Windows 11 Start menu designs it explored before settling on the least interesting one of the bunch

      May 13, 2025

      NVIDIA’s new GPU driver adds DOOM: The Dark Ages support and improves DLSS in Microsoft Flight Simulator 2024

      May 13, 2025

      How to install and use Ollama to run AI LLMs on your Windows 11 PC

      May 13, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Community News: Latest PECL Releases (05.13.2025)

      May 13, 2025
      Recent

      Community News: Latest PECL Releases (05.13.2025)

      May 13, 2025

      How We Use Epic Branches. Without Breaking Our Flow.

      May 13, 2025

      I think the ergonomics of generators is growing on me.

      May 13, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      This $4 Steam Deck game includes the most-played classics from my childhood — and it will save you paper

      May 13, 2025
      Recent

      This $4 Steam Deck game includes the most-played classics from my childhood — and it will save you paper

      May 13, 2025

      Microsoft shares rare look at radical Windows 11 Start menu designs it explored before settling on the least interesting one of the bunch

      May 13, 2025

      NVIDIA’s new GPU driver adds DOOM: The Dark Ages support and improves DLSS in Microsoft Flight Simulator 2024

      May 13, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»This AI Paper from Cohere for AI Presents a Comprehensive Study on Multilingual Preference Optimization

    This AI Paper from Cohere for AI Presents a Comprehensive Study on Multilingual Preference Optimization

    July 8, 2024

    Multilingual natural language processing (NLP) is a rapidly advancing field that aims to develop language models capable of understanding & generating text in multiple languages. These models facilitate effective communication and information access across diverse linguistic backgrounds. This field’s importance lies in its potential to bridge the gap between different language speakers, making technological advancements in AI accessible globally. However, developing such models presents significant challenges due to the complexities of handling multiple languages simultaneously.

    One of the main issues in multilingual NLP is the predominant focus on a few major languages, such as English and Chinese. This narrow concentration results in a significant performance gap for models when applied to less commonly spoken languages. Consequently, many languages still need to be represented, limiting AI technologies’ applicability and fairness. Addressing this disparity requires innovative approaches to enhance the quality and diversity of multilingual datasets, ensuring that AI models can perform effectively across a broad spectrum of languages.

    Traditional methods for improving multilingual language models often involve translating preference data from English to other languages. While this strategy helps somewhat, it introduces several problems, including translation artifacts that can degrade model performance. Relying heavily on translation can lead to a lack of diversity in the data, which is crucial for robust model training. Collecting high-quality multilingual preference data through human annotation is a potential solution, but it is both expensive and time-consuming, making it impractical for large-scale applications.

    Researchers from Cohere For AI have developed a novel, scalable method for generating high-quality multilingual feedback data. This method aims to balance data coverage and improve the performance of multilingual large language models (LLMs). The research team introduced a unique approach that leverages diverse, multilingual prompts and completions generated by multiple LLMs. This strategy not only increases the diversity of the data but also helps avoid the common pitfalls associated with translation artifacts. The models used in this research include Cohere’s Command and Command R+, specifically designed for multilingual capabilities.

    The methodology involves translating approximately 50,000 English prompts into 22 additional languages using the NLLB 3.3B model. These prompts are then used to generate completions in each language, ensuring high diversity and quality in the data. The research team also compared completions generated directly in the target language to those translated from English, finding that the former significantly reduced the occurrence of translation artifacts. This approach resulted in a diverse set of multilingual preference pairs crucial for effective preference optimization.

    The performance of the preference-trained model was evaluated against several state-of-the-art multilingual LLMs. The results were impressive, with the preference-trained model achieving a 54.4% win rate against Aya 23 8B, the current leading multilingual LLM in its parameter class. Additionally, the model showed a 69.5% win rate or higher against other widely used models such as Gemma-1.1-7B-it, Meta-Llama3-8B-Instruct, and Mistral-7B-Instruct-v0.3. These results highlight the effectiveness of the researchers’ approach in improving the performance of multilingual LLMs through enhanced preference optimization.

    Further analysis revealed that increasing the number of languages in the training data consistently improved the model’s performance. For example, training with five languages resulted in a win rate of 54.9% on unseen languages, compared to 46.3% when training only in English. Moreover, online preference optimization methods, such as Reinforcement Learning from Human Feedback (RLHF), proved more effective than offline methods like Direct Preference Optimization (DPO). The online techniques achieved higher win rates, with RLOO outperforming DPO by a margin of 10.6% in some cases.

    In conclusion, the research conducted by Cohere For AI demonstrates the critical importance of high-quality, diverse, multilingual data in training effective multilingual language models. The innovative methods introduced by the research team address the challenges of data scarcity and quality, resulting in performance improvements across a wide range of languages. The study not only sets a new benchmark for multilingual preference optimization but also underscores the value of online training methods in achieving superior cross-lingual transfer and overall model performance.

    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. 

    Join our Telegram Channel and LinkedIn Group.

    If you like our work, you will love our newsletter..

    Don’t Forget to join our 46k+ ML SubReddit

    The post This AI Paper from Cohere for AI Presents a Comprehensive Study on Multilingual Preference Optimization appeared first on MarkTechPost.

    Source: Read More 

    Hostinger
    Facebook Twitter Reddit Email Copy Link
    Previous ArticleGenerate unique images by fine-tuning Stable Diffusion XL with Amazon SageMaker
    Next Article MIA-Bench: Towards Better Instruction Following Evaluation of Multimodal LLMs

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 14, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-3623 – WordPress Uncanny Automator PHP Object Injection Vulnerability

    May 14, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    This AI Paper Introduces HalluVault for Detecting Fact-Conflicting Hallucinations in Large Language Models

    Development

    Join our Free Webinar: “Removing Friction from Design System Workflows”

    Development

    Fine-tune Meta Llama 3.2 text generation models for generative AI inference using Amazon SageMaker JumpStart

    Development

    Thailand Enhances Cyber Resilience with AI-Powered Security Measures

    Development

    Highlights

    Crawl, then walk, before you run with AI agents, experts recommend

    February 11, 2025

    Agentic AI offers compelling productivity benefits, but designers and developers must think small. Source: Latest…

    UBC Researchers Introduce ‘First Explore’: A Two-Policy Learning Approach to Rescue Meta-Reinforcement Learning RL from Failed Explorations

    December 17, 2024

    deepc: A Germany-based Radiology AI Startup that has Developed ​​the Leading AI Operating System for Radiologists

    August 1, 2024

    Best Free and Open Source Alternatives to Apple iMovie

    March 24, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.