Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      June 2, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      June 2, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      June 2, 2025

      How To Prevent WordPress SQL Injection Attacks

      June 2, 2025

      How Red Hat just quietly, radically transformed enterprise server Linux

      June 2, 2025

      OpenAI wants ChatGPT to be your ‘super assistant’ – what that means

      June 2, 2025

      The best Linux VPNs of 2025: Expert tested and reviewed

      June 2, 2025

      One of my favorite gaming PCs is 60% off right now

      June 2, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      `document.currentScript` is more useful than I thought.

      June 2, 2025
      Recent

      `document.currentScript` is more useful than I thought.

      June 2, 2025

      Adobe Sensei and GenAI in Practice for Enterprise CMS

      June 2, 2025

      Over The Air Updates for React Native Apps

      June 2, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      You can now open ChatGPT on Windows 11 with Win+C (if you change the Settings)

      June 2, 2025
      Recent

      You can now open ChatGPT on Windows 11 with Win+C (if you change the Settings)

      June 2, 2025

      Microsoft says Copilot can use location to change Outlook’s UI on Android

      June 2, 2025

      TempoMail — Command Line Temporary Email in Linux

      June 2, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»ByteDance Research Introduces 1.58-bit FLUX: A New AI Approach that Gets 99.5% of the Transformer Parameters Quantized to 1.58 bits

    ByteDance Research Introduces 1.58-bit FLUX: A New AI Approach that Gets 99.5% of the Transformer Parameters Quantized to 1.58 bits

    December 31, 2024

    Vision Transformers (ViTs) have become a cornerstone in computer vision, offering strong performance and adaptability. However, their large size and computational demands create challenges, particularly for deployment on devices with limited resources. Models like FLUX Vision Transformers, with billions of parameters, require substantial storage and memory, making them impractical for many use cases. These limitations restrict the real-world application of advanced generative models. Addressing these challenges calls for innovative methods to reduce the computational burden without compromising performance.

    Researchers from ByteDance Introduce 1.58-bit FLUX

    Researchers from ByteDance have introduced the 1.58-bit FLUX model, a quantized version of the FLUX Vision Transformer. This model reduces 99.5% of its parameters (11.9 billion in total) to 1.58 bits, significantly lowering computational and storage requirements. The process is unique in that it does not rely on image data, instead using a self-supervised approach based on the FLUX.1-dev model. By incorporating a custom kernel optimized for 1.58-bit operations, the researchers achieved a 7.7× reduction in storage and a 5.1× reduction in inference memory usage, making deployment in resource-constrained environments more feasible.

    Technical Details and Benefits

    The core of the 1.58-bit FLUX lies in its quantization technique, which restricts model weights to three values: +1, -1, or 0. This approach compresses parameters from 16-bit precision to 1.58 bits. Unlike traditional methods, this data-free quantization relies solely on a calibration dataset of text prompts, removing the need for image data. To handle the complexities of low-bit operations, a custom kernel was developed to optimize computations. These advances lead to substantial reductions in storage and memory requirements while maintaining the ability to generate high-resolution images of 1024 × 1024 pixels.

    Results and Insights

    Extensive evaluations of the 1.58-bit FLUX model on benchmarks such as GenEval and T2I CompBench demonstrated its efficacy. The model delivered performance on par with its full-precision counterpart, with minor deviations observed in specific tasks. In terms of efficiency, the model achieved a 7.7× reduction in storage and a 5.1× reduction in memory usage across various GPUs. Deployment-friendly GPUs, such as the L20 and A10, further highlighted the model’s practicality with notable latency improvements. These results indicate that 1.58-bit FLUX effectively balances efficiency and performance, making it suitable for a range of applications.

    Conclusion

    The development of 1.58-bit FLUX addresses critical challenges in deploying large-scale Vision Transformers. Its ability to significantly reduce storage and memory requirements without sacrificing performance represents a step forward in efficient AI model design. While there is room for improvement, such as enhancing activation quantization and fine-detail rendering, this work sets a solid foundation for future advancements. As research continues, the prospect of deploying high-quality generative models on everyday devices becomes increasingly realistic, broadening access to powerful AI capabilities.


    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

    🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

    The post ByteDance Research Introduces 1.58-bit FLUX: A New AI Approach that Gets 99.5% of the Transformer Parameters Quantized to 1.58 bits appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleMaximizing ROI with Custom Software Solutions
    Next Article Revolutionizing LLM Alignment: A Deep Dive into Direct Q-Function Optimization

    Related Posts

    Security

    Chrome Zero-Day Alert: CVE-2025-5419 Actively Exploited in the Wild

    June 2, 2025
    Security

    CISA Adds 5 Actively Exploited Vulnerabilities to KEV Catalog: ASUS Routers, Craft CMS, and ConnectWise Targeted

    June 2, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Tap into Your PHP Potential with Free Projects at PHPGurukul

    Development

    Beyond Quadratic Bottlenecks: Mamba-2 and the State Space Duality Framework for Efficient Language Modeling

    Development

    How Formula 1® uses generative AI to accelerate race-day issue resolution

    Machine Learning

    Code Your Own Llama 4 LLM from Scratch

    Development

    Highlights

    CVE-2024-13381 – WordPress Calculated Fields Form Stored Cross-Site Scripting Vulnerability

    May 1, 2025

    CVE ID : CVE-2024-13381

    Published : May 1, 2025, 6:15 a.m. | 1 hour, 56 minutes ago

    Description : The Calculated Fields Form WordPress plugin before 5.2.62 does not sanitise and escape some of its settings, which could allow high privilege users such as admin to perform Stored Cross-Site Scripting attacks even when the unfiltered_html capability is disallowed (for example in multisite setup).

    Severity: 0.0 | NA

    Visit the link for more details, such as CVSS details, affected products, timeline, and more…

    Top Figma Competitors

    August 2, 2024

    Mastering Page Properties With Granite Render Conditions and Context-Aware Configuration

    July 29, 2024

    Integrate generative AI capabilities into Microsoft Office using Amazon Bedrock

    March 19, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.