Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      May 16, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      May 16, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      May 16, 2025

      How To Prevent WordPress SQL Injection Attacks

      May 16, 2025

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025

      Minecraft licensing robbed us of this controversial NFL schedule release video

      May 16, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      The power of generators

      May 16, 2025
      Recent

      The power of generators

      May 16, 2025

      Simplify Factory Associations with Laravel’s UseFactory Attribute

      May 16, 2025

      This Week in Laravel: React Native, PhpStorm Junie, and more

      May 16, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025
      Recent

      Microsoft has closed its “Experience Center” store in Sydney, Australia — as it ramps up a continued digital growth campaign

      May 16, 2025

      Bing Search APIs to be “decommissioned completely” as Microsoft urges developers to use its Azure agentic AI alternative

      May 16, 2025

      Microsoft might kill the Surface Laptop Studio as production is quietly halted

      May 16, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»This AI Paper by National University of Singapore Introduces A Comprehensive Survey of Language Models for Tabular Data Analysis

    This AI Paper by National University of Singapore Introduces A Comprehensive Survey of Language Models for Tabular Data Analysis

    August 23, 2024

    Tabular data, which dominates many genres, such as healthcare, financial, and social science applications, contains rows and columns with structured features, making it much easier for data management or analysis. However, the diversity of tabular data, including numerical, unconditional, and textual, brings huge challenges to attaining robust and accurate predictive performance. Another area for improvement in effectively modeling and analyzing this type of data is the complexity of the relationships inside the data, particularly dependencies between rows and columns.

    The main challenge in analyzing tabular data is that it is very difficult to handle its heterogeneous structure. Traditional machine learning models stay far away when considering the complex relationships inside tabular datasets, especially for large and complex datasets. These models require additional guidance to generalize well in the presence of a diversity of data types and interdependencies of tabular data. This challenge becomes even more complex given the need for high predictive accuracy and robustness, especially in critical applications such as health care, where the decisions among data analysis can be quite consequential.

    Different methods have been applied to overcome these challenges of modeling tabular data. Early techniques relied largely on conventional machine learning, most of which needed a lot of feature engineering to model the subtleties of the data. The known weakness of these naturally lay in their inability to scale in size and complexity of the input dataset. More recently, techniques from NLP have been adapted for tabular data; more specifically, transformer-based architectures are increasingly implemented. These methods started by training the transformers from scratch over tabular data, but this had the disadvantage of needing huge amounts of training data with significant scalability issues. Against this backdrop, researchers began using PLMs like BERT, which required less data and provided better predictive performance.

    Researchers from the National University of Singapore provided a comprehensive survey of the various language modeling techniques developed for tabular data. The review systematizes classification for literature and further identifies a trend shift from traditional machine learning models to advanced methods using state-of-the-art LLMs like GPT and LLaMA. This research has emphasized the evolution of these models, showing how LLMs have been radical in the field, taking it further into more sophisticated applications in modeling tabular data. This work is important to fill a gap in the relevant literature by providing a detailed taxonomy of the tabular data structures, key datasets, and various modeling techniques.

    The methodology proposed by the research team categorizes tabular data into two major categories: 1D and 2D. On the other hand, 1D tabular data usually contains only one table, with the main work coming at the row level, which, of course, is simpler but very important for tasks like classification and regression. In contrast, 2D tabular data consists of multiple related tables, requiring more complex modeling techniques for tasks such as table retrieval and table question-answering. The researchers delve into different strategies for turning tabular data into forms that their language model can consume. These strategies include flattening sequences, row processing, and integrating this information in prompts. Through these methods, the language models lever a more profound understanding and processing abilities of tabular data towards assured predictive outcomes.

    The research shows how strong the ability of great big language models is in most tasks of tabular data. These models have demonstrated marked improvement in understanding and processing complex data structures on functions such as Table Question Answering and Table Semantic Parsing. The authors illustrate how LLMs enable a standard rise in all tasks at higher levels of accuracy and efficiency by exploiting pre-trained knowledge and advanced attention mechanisms that set new tabular data modeling standards across many applications.

    In conclusion, the research has underscored the potential that NLP techniques have for effectively changing the very nature of tabular data analysis in the presence of large language models. By systematizing the review and categorization of existing methods, researchers have proposed a very clear roadmap for future developments in this area. The proposed methodologies negate the intrinsic challenges of tabular data and open up new advanced applications with guarantees of relevance and effectiveness, including when the complexity of data rises.

    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

    Don’t Forget to join our 49k+ ML SubReddit

    Find Upcoming AI Webinars here

    The post This AI Paper by National University of Singapore Introduces A Comprehensive Survey of Language Models for Tabular Data Analysis appeared first on MarkTechPost.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleProcessing 2-Hour Videos Seamlessly: This AI Paper Unveils LONGVILA, Advancing Long-Context Visual Language Models for Long Videos
    Next Article DeepSim: AI-Accelerated 3D Physics Simulator for Engineers

    Related Posts

    Security

    Nmap 7.96 Launches with Lightning-Fast DNS and 612 Scripts

    May 16, 2025
    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-47916 – Invision Community Themeeditor Remote Code Execution

    May 16, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Optimizing Spiking Neural P Systems Simulations: Achieving Unprecedented Speed and Efficiency through Compressed Matrix Representations on GPUs Using CUDA

    Development

    Qual è la Differenza tra un PC e un Server?

    Linux

    DeepSeek vs ChatGPT: The Ultimate AI Model Comparison

    Development

    Google: Over 57 Nation-State Threat Groups Using AI for Cyber Operations

    Development
    GetResponse

    Highlights

    I want to write Java code in Jmeter to copy current date in a variable and use this variable in next request. Where should I write it?

    August 3, 2024

    This is my code:

    package dateTime;
    import java.text.DateFormat;
    import java.text.SimpleDateFormat;
    import java.time.LocalDate;
    import java.time.LocalDateTime;
    import java.time.format.DateTimeFormatter;
    import java.util.Calendar;
    import java.util.Date;

    public class GetCurrentDateTime {

    public static void main(String[] args) {

    LocalDate localDate = LocalDate.now();
    String date = DateTimeFormatter.ofPattern(“yyyy/MM/dd”).format(localDate);
    log.info(“The date is “+date);
    vars.put(“currentDate”,date);

    }

    }

    I tried above code in beanshell, JSR223 pre-processor or post processor but not able to use it further.

    How to maximize browser window irrespective of windows versions

    June 4, 2024

    I biked with these photochromic glasses, and the heads-up display was game-changing

    August 30, 2024

    DAI#43 – AI Apples, search power, and Meta’s EU ‘pretty please’

    June 14, 2024
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.