Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 16, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 16, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 16, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 16, 2025

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025

      Minecraft licensing robbed us of this controversial NFL schedule release video

      May 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      The power of generators

      May 16, 2025
      Recent

      The power of generators

      May 16, 2025

      Simplify Factory Associations with Laravel’s UseFactory Attribute

      May 16, 2025

      This Week in Laravel: React Native, PhpStorm Junie, and more

      May 16, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025
      Recent

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Researchers at the University of Freiburg and Bosch AI Propose HW-GPT-Bench: A Hardware-Aware Language Model Surrogate Benchmark

    Researchers at the University of Freiburg and Bosch AI Propose HW-GPT-Bench: A Hardware-Aware Language Model Surrogate Benchmark

    May 23, 2024

    Large language models (LLMs) have recently become highly valuable tools in complicated reasoning tasks, language production, and human language interpretation. Since then, there has been a dramatic increase in funding for studies in this area, and both the number of models used and the amount of data used for training have grown substantially. This also points to a rise in the inference and training costs. 

    Having efficient designs at inference time is important to ensure these models’ broader range of uses and flexibility. Various Pareto-Frontiers, or trade-offs, between LLM latency and performance, are relevant to these systems’ end-users. Multiple strategies, such as pruning and KV-Cache optimization, have been used to improve the inference efficiency of language models. Finding the best frontier of language models for inference can thus be expressed as a problem of optimizing many objectives or constraints.

    A new study by researchers from the University of Freiburg and Bosch Center for Artificial Intelligence present Hardware-Aware-GPT-Bench (HW-GPT-Bench), a language model space benchmark that takes hardware into account, to evaluate and optimize LLMs (long language models) using various hardware metrics. The goal of creating this benchmark is to speed up the process of studying and developing algorithms for hardware-aware search in the language model space.

    To efficiently train a supernet proxy that covers different LLM setups, HW-GPT-Bench uses weight-sharing methods from Neural Architecture Search (NAS). A complete evaluation methodology is provided by profiling these models on thirteen devices using five critical hardware metrics: latency, energy consumption, GPU memory usage, FLOPS, and performance. 

    This comprehensive benchmark covers small, medium, and large model scales using performance and hardware metric predictors across many devices. The team investigated eight distinct multi-objective optimization algorithms, comparing performance and hardware measurements to find the best configurations by analyzing cutting-edge NAS methods. They use their pretrained surrogates for various model sizes to investigate the interplay between hardware and performance measures. This work helps with integration and reproducibility; the public API provides a queryable, open-source interface for predictors, supernetwork weights, and baselines.

    Training and deploying LLMs place a heavy computational burden on the world’s power grid. To minimize the negative environmental effects caused by large-scale AI deployments, HW-GPT-Bench optimizes LLM configurations to lower energy consumption. The proposed benchmark helps create environmentally friendly AI by locating designs that use less power.

    Optimizing hardware efficiency during LLMs’ training and deployment stages can result in significant cost savings. By decreasing the computational resources required, organizations can reap economic benefits and make large-scale AI solution deployment more realistic. Industries that rely on processing and analyzing massive amounts of data will benefit the most from this economic efficiency.

    The team’s long-term goals include:

    Investigating quantization methods.

    Developing surrogates for more current and larger models.

    Determining the best way to combine NAS with pruning strategies.

    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

    If you like our work, you will love our newsletter..

    Don’t Forget to join our 42k+ ML SubReddit

    The post Researchers at the University of Freiburg and Bosch AI Propose HW-GPT-Bench: A Hardware-Aware Language Model Surrogate Benchmark appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleOpenRLHF: An Open-Source AI Framework Enabling Efficient Reinforcement Learning from Human Feedback RLHF Scaling
    Next Article Demystifying Vision-Language Models: An In-Depth Exploration

    Related Posts

    Machine Learning

    LLMs Struggle with Real Conversations: Microsoft and Salesforce Researchers Reveal a 39% Performance Drop in Multi-Turn Underspecified Tasks

    May 17, 2025
    Machine Learning

    This AI paper from DeepSeek-AI Explores How DeepSeek-V3 Delivers High-Performance Language Modeling by Minimizing Hardware Overhead and Maximizing Computational Efficiency

    May 17, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Unable to scroll element in Appium 1.20

    Development

    The Noodle Bus Bash

    Artificial Intelligence

    New game blending Valheim and Mount & Blade comes to Xbox later this year

    News & Updates

    CVE-2025-4135 – Netgear WG302v2 Command Injection Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    Highlights

    Development

    Cybercriminals Send Fake Legal Notices to Scam Indians: Here’s What to Know

    December 27, 2024

    The authorities of India have issued an urgent warning to citizens regarding the circulation of…

    Moxa Issues Fix for Critical Authentication Bypass Vulnerability in PT Switches

    March 16, 2025

    Server-side Rescoring of Spoken Entity-centric Knowledge Queries for Virtual Assistants

    June 20, 2024

    The developers behind Chrono Trigger and Final Fantasy XV will deliver another JRPG sequel to Xbox and PC gamers this year

    February 26, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.