Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      The state of DevOps and AI: Not just hype

      September 1, 2025

      A Breeze Of Inspiration In September (2025 Wallpapers Edition)

      August 31, 2025

      10 Top Generative AI Development Companies for Enterprise Node.js Projects

      August 30, 2025

      Prompting Is A Design Act: How To Brief, Guide And Iterate With AI

      August 29, 2025

      Look out, Meta Ray-Bans! These AI glasses just raised over $1M in pre-orders in 3 days

      September 2, 2025

      Samsung ‘Galaxy Glasses’ powered by Android XR are reportedly on track to be unveiled this month

      September 2, 2025

      The M4 iPad Pro is discounted $100 as a last-minute Labor Day deal

      September 2, 2025

      Distribution Release: Linux From Scratch 12.4

      September 1, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Enhanced Queue Job Control with Laravel’s ThrottlesExceptions failWhen() Method

      September 2, 2025
      Recent

      Enhanced Queue Job Control with Laravel’s ThrottlesExceptions failWhen() Method

      September 2, 2025

      August report 2025

      September 2, 2025

      Fake News Detection using Python Machine Learning (ML)

      September 1, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Installing Proxmox on a Raspberry Pi to run Virtual Machines on it

      September 2, 2025
      Recent

      Installing Proxmox on a Raspberry Pi to run Virtual Machines on it

      September 2, 2025

      Download Transcribe! for Windows

      September 1, 2025

      Microsoft Fixes CertificateServicesClient (CertEnroll) Error in Windows 11

      September 1, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»Moonsight AI Released Kimi-VL: A Compact and Powerful Vision-Language Model Series Redefining Multimodal Reasoning, Long-Context Understanding, and High-Resolution Visual Processing

    Moonsight AI Released Kimi-VL: A Compact and Powerful Vision-Language Model Series Redefining Multimodal Reasoning, Long-Context Understanding, and High-Resolution Visual Processing

    April 12, 2025

    Multimodal AI enables machines to process and reason across various input formats, such as images, text, videos, and complex documents. This domain has seen increased interest as traditional language models, while powerful, are inadequate when confronted with visual data or when contextual interpretation spans across multiple input types. The real world is inherently multimodal, so systems aiming to assist in real-time tasks, analyzing user interfaces, understanding academic materials, or interpreting complex scenes require intelligence that functions beyond textual reasoning. Newer models are now being developed to simultaneously decode language and vision cues to approach tasks with improved contextual awareness, reasoning depth, and adaptability to different data input forms.

    A limitation in multimodal systems today lies in their inability to process long contexts efficiently and to generalize across high-resolution or diverse input structures without compromising performance. Many open-source models limit the input to a few thousand tokens or demand excessive computational resources to maintain performance at scale. These constraints result in models that may perform well on standard benchmarks but struggle with real-world applications that involve complex, multi-image inputs, extended dialogues, or academic tasks like OCR-based document analysis and mathematical problem-solving. There’s also a gap in reasoning ability, particularly long-horizon thinking, which prevents current systems from handling tasks that require step-by-step logic or deep contextual alignment between different data modalities.

    Previous tools have attempted to address these challenges but often fell short in scalability or flexibility. The Qwen2.5-VL series and Gemma-3 models, while notable for their dense architectures, lack built-in support for reasoning through longer chains of thought. Models like DeepSeek-VL2 and Aria adopted mixture-of-experts (MoE) strategies but had fixed vision encoders that restricted their ability to adapt to various resolutions and forms of visual input. Also, these models typically supported only short context windows, 4K tokens in DeepSeek-VL2, and had limited success in complex OCR or multi-image scenarios. As such, most existing systems failed to balance low resource consumption with the ability to tackle tasks involving long context and diverse visual data.

    Researchers at Moonshot AI introduced Kimi-VL, a novel vision-language model utilizing an MoE architecture. This system activates only 2.8 billion parameters in its decoder, significantly lighter than many competitors while maintaining powerful multimodal capabilities. The two released models based on this architecture on Hugging Face are Kimi-VL-A3B-Thinking and Kimi-VL-A3B-Instruct. It incorporates a native-resolution visual encoder named MoonViT and supports context windows of up to 128K tokens. The model has three integrated components: the MoonViT encoder, an MLP projector for transitioning visual features to language embeddings, and the Moonlight MoE decoder. Researchers further developed an advanced version, Kimi-VL-Thinking, designed specifically for long-horizon reasoning tasks through chain-of-thought supervised fine-tuning and reinforcement learning. Together, these models aim to redefine efficiency benchmarks in vision-language reasoning.

    The architectural innovation in Kimi-VL lies in its adaptability and processing capability. MoonViT processes high-resolution images in their original form, eliminating the need for sub-image fragmentation. To ensure spatial consistency across varied image resolutions, the model uses interpolated absolute positional embeddings combined with two-dimensional rotary positional embeddings across both height and width. These design choices allow MoonViT to preserve fine-grained detail even in large-scale image inputs. Outputs from the vision encoder are passed through a two-layer MLP that uses pixel shuffle operations to downsample spatial dimensions and convert features into LLM-compatible embeddings. On the language side, the 2.8B activated parameter MoE decoder supports 16B total parameters and integrates seamlessly with visual representations, enabling highly efficient training and inference across different input types. The entire training process used an enhanced Muon optimizer with weight decay and ZeRO-1-based memory optimization for handling the large parameter count.

    The training data composition reflects a focus on diverse multimodal learning. Starting with 2.0T tokens for ViT training using image-caption pairs, the team added another 0.1T to align the encoder with the decoder. Joint pre-training consumed 1.4T tokens, followed by 0.6T in cooldown and 0.3T in long-context activation, totaling 4.4T tokens. These stages included academic visual datasets, OCR samples, long video data, and synthetic mathematical and code-based QA pairs. For long-context learning, the model was progressively trained to handle sequences from 8K up to 128K tokens, using RoPE embeddings extended from a base frequency of 50,000 to 800,000. This allowed the model to maintain a token recall accuracy of 100% up to 64K tokens, with a slight drop to 87.0% at 128K, still outperforming most alternatives.

    Kimi-VL demonstrated strong results across a range of benchmarks. On the LongVideoBench, it scored 64.5; on MMLongBench-Doc, it achieved 35.1; and on the InfoVQA benchmark, it led with 83.2. On ScreenSpot-Pro, which tests understanding of UI screens, it scored 34.5. The Kimi-VL-Thinking variant excelled in reasoning-intensive benchmarks like MMMU (61.7), MathVision (36.8), and MathVista (71.3). For agent tasks such as OSWorld, the model matched or exceeded performance from larger models like GPT-4o while activating significantly fewer parameters. Its compact design and strong reasoning capabilities make it a leading candidate among open-source multimodal solutions.

    Some Key Takeaways from the Research on Kimi-VL:

    • Kimi-VL activates only 2.8B parameters during inference, ensuring efficiency without sacrificing capability. 
    • MoonViT, its vision encoder, natively processes high-resolution images, improving clarity in tasks like OCR and UI interpretation.  
    • The model supports up to 128K context tokens, achieving 100% recall up to 64K and 87.0% accuracy at 128K on text/video tasks.  
    • Kimi-VL-Thinking scores 61.7 on MMMU, 36.8 on MathVision, and 71.3 on MathVista, outperforming many larger VLMs.  
    • It scored 83.2 on InfoVQA and 34.5 on visual tasks on ScreenSpot-Pro, showcasing its precision in perception-based evaluations.  
    • Total pre-training involved 4.4T tokens across text, video, document, and synthetic multimodal data. 
    • Optimization was done using a customized Muon optimizer with memory-efficient strategies like ZeRO-1. 
    • Joint training ensured seamless visual and language feature integration while preserving core language capabilities.

    Check out Instruct Model and Reasoning Model. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 85k+ ML SubReddit.

    The post Moonsight AI Released Kimi-VL: A Compact and Powerful Vision-Language Model Series Redefining Multimodal Reasoning, Long-Context Understanding, and High-Resolution Visual Processing appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleStep by Step Coding Guide to Build a Neural Collaborative Filtering (NCF) Recommendation System with PyTorch
    Next Article DuckDB is an in-process SQL OLAP database management system

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    September 2, 2025
    Machine Learning

    Introducing auto scaling on Amazon SageMaker HyperPod

    August 30, 2025
    Leave A Reply Cancel Reply

    For security, use of Google's reCAPTCHA service is required which is subject to the Google Privacy Policy and Terms of Use.

    Continue Reading

    Intel Plans ‘Arrow Lake Refresh’ With Faster Clock and New NPU This Year

    Operating Systems

    CVE-2025-48739 – TheHive SSRF

    Common Vulnerabilities and Exposures (CVEs)

    Distribution Release: AlmaLinux OS 10.0

    News & Updates

    HazyBeacon: Novel Backdoor Uses AWS Lambda for Stealthy C2, Targets Govts

    Security

    Highlights

    CVE-2025-32404 – RT-Labs P-Net OOB Write Vulnerability

    May 7, 2025

    CVE ID : CVE-2025-32404

    Published : May 7, 2025, 7:15 a.m. | 20 minutes ago

    Description : An Out-of-bounds Write in RT-Labs P-Net version 1.0.1 or earlier allows an attacker to corrupt the memory of IO devices that use the library by sending a malicious RPC packet.

    Severity: 4.8 | MEDIUM

    Visit the link for more details, such as CVSS details, affected products, timeline, and more…

    AI-Driven Automation in 2025: A Game-Changer for Small Business Efficiency⚙️

    June 24, 2025

    PoC Exploit Released for Apache Tomcat DoS Vulnerability

    June 6, 2025

    Red Hat Unveils llm-d: Scaling Generative AI for the Enterprise

    June 2, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.