Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 12, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 12, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 12, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 12, 2025

      Microsoft aims to be “carbon negative” by 2030, with 3 million carbon removal credits in its backyard of Washington

      May 12, 2025

      Sam Altman doesn’t want his son to have an AI “bestie” — as Microsoft plans to turn Copilot into an AI friend and companion

      May 12, 2025

      ChatGPT downplays AI’s threat to humanity despite an apparent “99.999999% probability” of inevitable doom

      May 12, 2025

      Surface Pro 12-inch vs. iPad Air M3: Which should you choose?

      May 12, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      A customizable and accessible web component

      May 12, 2025
      Recent

      A customizable and accessible web component

      May 12, 2025

      How Agile Helps You Improve Your Agility

      May 12, 2025

      Laravel Seeder Generator

      May 12, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft aims to be “carbon negative” by 2030, with 3 million carbon removal credits in its backyard of Washington

      May 12, 2025
      Recent

      Microsoft aims to be “carbon negative” by 2030, with 3 million carbon removal credits in its backyard of Washington

      May 12, 2025

      Sam Altman doesn’t want his son to have an AI “bestie” — as Microsoft plans to turn Copilot into an AI friend and companion

      May 12, 2025

      ChatGPT downplays AI’s threat to humanity despite an apparent “99.999999% probability” of inevitable doom

      May 12, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Meet Tsinghua University’s GLM-4-9B-Chat-1M: An Outstanding Language Model Challenging GPT 4V, Gemini Pro (on vision), Mistral and Llama 3 8B

    Meet Tsinghua University’s GLM-4-9B-Chat-1M: An Outstanding Language Model Challenging GPT 4V, Gemini Pro (on vision), Mistral and Llama 3 8B

    June 6, 2024

    Tsinghua University’s Knowledge Engineering Group (KEG) has unveiled GLM-4 9B, a powerful new language model that outperforms GPT-4 and Gemini in various benchmarks. Developed by the Tsinghua Deep Model (THUDM) team, this open-source model marks a significant milestone in the field of natural language processing.

    At its core, GLM-4 9B is a massive language model trained on an unprecedented 10 trillion tokens spanning 26 languages. It caters to various capabilities, including multi-round dialogue in Chinese and English, code execution, web browsing, and custom tool calling through Function Call.

    The model’s architecture is built upon the latest advancements in deep learning, incorporating cutting-edge techniques such as attention mechanisms and transformer architectures. The base version supports a context window of up to 128,000 tokens, while a specialized variant allows for an impressive 1 million token context length.

    Compared to industry giants like GPT and Gemini, GLM-4 9B’s architecture stands out with its support for high-resolution vision tasks (up to 1198 x 1198 pixels) and its ability to handle a diverse range of languages. This versatility positions GLM-4 9B as a powerful contender in the language model landscape.

    Evaluations on various datasets have demonstrated GLM-4 9B’s superior performance in many areas and performance on par with the best models for some of the tasks, the model has surpassed every other existing model on overall accuracy. Notably, it has outperformed GPT-4, Gemini Pro (in vision tasks), Mistral, and Llama 3 8B, solidifying its position as a formidable force in the field.

    With its open-source nature and permissive commercial use (under certain conditions), GLM-4 9B presents a wealth of opportunities for developers, researchers, and businesses alike. Potential applications range from natural language processing tasks to computer vision, code generation, and beyond. The model’s seamless integration with the Transformers library further simplifies its adoption and deployment.

    The release of GLM-4 9B by Tsinghua University’s KEG marks a significant milestone in language models. With its impressive performance, multi-lingual capabilities, and versatile architecture, this model sets a new benchmark for open-source language models and paves the way for further advancements in natural language processing and artificial intelligence.

    Check out the Model on HF Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

    If you like our work, you will love our newsletter..

    Don’t Forget to join our 43k+ ML SubReddit | Also, check out our AI Events Platform

    The post Meet Tsinghua University’s GLM-4-9B-Chat-1M: An Outstanding Language Model Challenging GPT 4V, Gemini Pro (on vision), Mistral and Llama 3 8B appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleGROKFAST: A Machine Learning Approach that Accelerates Grokking by Amplifying Slow Gradients
    Next Article The Ultimate Guide to Healthcare SEO: Get Found by Patients

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 13, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-47858 – Apache HTTP Server Cross-Site Request Forgery

    May 13, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Form Data Extraction: From OCR to Deep Learning

    Artificial Intelligence

    Helping MongoDB Customers Unlock Potential with Industry Solutions

    Databases

    Ultralytics AI Library Compromised: Cryptocurrency Miner Found in PyPI Versions

    Development

    Consistency Large Language Models (CLLMs): A New Family of LLMs Specialized for the Jacobi Decoding Method for Latency Reduction

    Development
    GetResponse

    Highlights

    Development

    Meta AI Introduces Meta LLM Compiler: A State-of-the-Art LLM that Builds upon Code Llama with Improved Performance for Code Optimization and Compiler Reasoning

    June 28, 2024

    Software engineering has witnessed remarkable advancements with the development of Large Language Models (LLMs). These…

    New Cyble Report Highlights Critical Vulnerabilities and Rising Cyber Threats in ANZ for 2024

    November 21, 2024

    ChatGPT won’t interrupt you that often anymore, according to OpenAI

    March 26, 2025

    New Snake Keylogger Variant Leverages AutoIt Scripting to Evade Detection

    February 19, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.