Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 16, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 16, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 16, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 16, 2025

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025

      Minecraft licensing robbed us of this controversial NFL schedule release video

      May 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      The power of generators

      May 16, 2025
      Recent

      The power of generators

      May 16, 2025

      Simplify Factory Associations with Laravel’s UseFactory Attribute

      May 16, 2025

      This Week in Laravel: React Native, PhpStorm Junie, and more

      May 16, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025
      Recent

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»This AI Paper Introduces TabM: An Efficient Ensemble-Based Deep Learning Model for Robust Tabular Data Processing

    This AI Paper Introduces TabM: An Efficient Ensemble-Based Deep Learning Model for Robust Tabular Data Processing

    November 14, 2024

    By processing complex data formats, deep learning has transformed various domains, including finance, healthcare, and e-commerce. However, applying deep learning models to tabular data, characterized by rows and columns, poses unique challenges. While deep learning has excelled in image and text analysis, classic machine learning techniques such as gradient-boosted decision trees still dominate tabular data due to their reliability and interpretability. Researchers are exploring new architectures that can effectively adapt deep learning techniques for tabular data without sacrificing accuracy or efficiency.

    One significant challenge in applying deep learning to tabular data is balancing model complexity and computational efficiency. Traditional machine learning methods, particularly gradient-boosted decision trees, deliver consistent performance across diverse datasets. In contrast, deep learning models suffer from overfitting and require extensive computational resources, making them less practical for many real-world datasets. Furthermore, tabular data exhibits varied structures and distributions, making it challenging for deep learning models to generalize well. Thus, the need arises for a model that achieves high accuracy and remains efficient across diverse datasets.

    Current methods for tabular data in deep learning include multilayer perceptrons (MLPs), transformers, and retrieval-based models. While MLPs are simple and computationally light, they often fail to capture complex interactions within tabular data. More advanced architectures like transformers and retrieval-based methods introduce mechanisms such as attention layers to enhance feature interaction. However, these approaches often require significant computational resources, making them impractical for large datasets and limiting their widespread application. This gap in deep learning for tabular data led to exploring alternative, more efficient architectures.

    Researchers from Yandex and HSE University introduced a model named TabM, built upon an MLP foundation but enhanced with BatchEnsemble for parameter-efficient ensembling. This model generates multiple predictions within a single structure by sharing most of its weights among ensemble members, allowing it to produce diverse, weakly correlated predictions. By combining simplicity with effective ensembling, TabM balances efficiency and performance, aiming to outperform traditional MLP models without the complexity of transformer architectures. TabM offers a practical solution, providing advantages for deep learning without the excessive resource demands typically associated with advanced models.

    The methodology behind TabM leverages BatchEnsemble to maximize prediction diversity and accuracy while maintaining computational efficiency. Each ensemble member uses unique weights, known as adapters, to create a range of predictions. TabM generates robust outputs by averaging these predictions, mitigating overfitting, and improving generalization across diverse datasets. The researchers’ approach combines MLP simplicity with efficient ensembling, creating a balanced model architecture that enhances predictive accuracy and is less prone to common tabular data pitfalls. TabM’s efficient design allows it to achieve high accuracy on complex datasets without the heavy computational demands of transformer-based methods.

    Empirical evaluations demonstrate TabM’s strong performance across 46 public datasets, showing an average improvement of approximately 2.07% over standard MLP models. Specifically, on domain-aware splits—representing more complex, real-world scenarios—TabM outperformed many other deep learning models, proving its robustness. TabM showcased efficient processing capabilities on large datasets, managing datasets with up to 6.5 million objects on the Maps Routing dataset within 15 minutes. For classification tasks, TabM utilized the ROC-AUC metric, achieving consistent accuracy. At the same time, Root Mean Squared Error (RMSE) was employed for regression tasks, demonstrating the model’s capacity to generalize effectively across various task types.

    The study presents a significant advancement in applying deep learning to tabular data, merging MLP efficiency with an innovative ensembling strategy that optimizes computational demands and accuracy. By addressing the limitations of previous models, TabM provides an accessible and reliable solution that meets the needs of practitioners handling diverse tabular data types. As an alternative to traditional gradient-boosted decision trees and complex neural architectures, TabM represents a valuable development, offering a streamlined, high-performing model capable of efficiently processing real-world tabular datasets.


    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 55k+ ML SubReddit.

    [FREE AI WEBINAR] Implementing Intelligent Document Processing with GenAI in Financial Services and Real Estate Transactions

    The post This AI Paper Introduces TabM: An Efficient Ensemble-Based Deep Learning Model for Robust Tabular Data Processing appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleIntroducing Stable Diffusion 3.5 Large in Amazon SageMaker JumpStart
    Next Article No Train, All Gain: Enhancing Deep Frozen Representations with Self-Supervised Gradients

    Related Posts

    Machine Learning

    LLMs Struggle with Real Conversations: Microsoft and Salesforce Researchers Reveal a 39% Performance Drop in Multi-Turn Underspecified Tasks

    May 17, 2025
    Machine Learning

    This AI paper from DeepSeek-AI Explores How DeepSeek-V3 Delivers High-Performance Language Modeling by Minimizing Hardware Overhead and Maximizing Computational Efficiency

    May 17, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Understanding the Language Server Protocol – Easier Code Editing Across Languages and Tools

    Development

    How to Automate Mobile Testing: Strategies for Reliable, Scalable Tests

    Development

    Google AI Released Gemini 2.5 Pro Experimental: An Advanced AI Model that Excels in Reasoning, Coding, and Multimodal Capabilities

    Machine Learning

    The Founder’s Guide to Mastering Design

    Development
    GetResponse

    Highlights

    Development

    Commando Cat Cryptojacking Attacks Target Misconfigured Docker Instances

    June 7, 2024

    The threat actor known as Commando Cat has been linked to an ongoing cryptojacking attack…

    Kodeco Podcast: Surviving Layoffs With Joey deVilla (V2, S2, E11) [FREE]

    June 13, 2024

    Sam Altman wants OpenAI to be the Microsoft of AI, with a subscription-based operating system built on ChatGPT

    May 15, 2025

    Why is Microsoft firing its employees—again? Third round of layoff is happening

    July 5, 2024
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.