Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Sunshine And March Vibes (2025 Wallpapers Edition)

      June 1, 2025

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      June 1, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      June 1, 2025

      How To Prevent WordPress SQL Injection Attacks

      June 1, 2025

      7 MagSafe accessories that I recommend every iPhone user should have

      June 1, 2025

      I replaced my Kindle with an iPad Mini as my ebook reader – 8 reasons why I don’t regret it

      June 1, 2025

      Windows 11 version 25H2: Everything you need to know about Microsoft’s next OS release

      May 31, 2025

      Elden Ring Nightreign already has a duos Seamless Co-op mod from the creator of the beloved original, and it’ll be “expanded on in the future”

      May 31, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Student Record Android App using SQLite

      June 1, 2025
      Recent

      Student Record Android App using SQLite

      June 1, 2025

      When Array uses less memory than Uint8Array (in V8)

      June 1, 2025

      Laravel 12 Starter Kits: Definite Guide Which to Choose

      June 1, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Photobooth is photobooth software for the Raspberry Pi and PC

      June 1, 2025
      Recent

      Photobooth is photobooth software for the Raspberry Pi and PC

      June 1, 2025

      Le notizie minori del mondo GNU/Linux e dintorni della settimana nr 22/2025

      June 1, 2025

      Rilasciata PorteuX 2.1: Novità e Approfondimenti sulla Distribuzione GNU/Linux Portatile Basata su Slackware

      June 1, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»University of South Florida Researchers Propose TeLU Activation Function for Fast and Stable Deep Learning

    University of South Florida Researchers Propose TeLU Activation Function for Fast and Stable Deep Learning

    January 3, 2025

    Inspired by the brain, neural networks are essential for recognizing images and processing language. These networks rely on activation functions, which enable them to learn complex patterns. However, many activation functions face challenges. Some struggle with vanishing gradients, which slows learning in deep networks, while others suffer from “dead neurons,” where certain parts of the network stop learning. Modern alternatives aim to solve these issues but often have drawbacks like inefficiency or inconsistent performance. 

    Currently, activation functions in neural networks face significant issues. Functions like step and sigmoid struggle with vanishing gradients, limiting their effectiveness in deep networks, and while tanh improved this slightly, which proved to have other issues. ReLU addresses some gradient problems but introduces the “dying ReLU” issue, making neurons inactive. Variants like Leaky ReLU and PReLU attempt fixes but bring inconsistencies and challenges in regularization. Advanced functions like ELU, SiLU, and GELU improve non-linearities. However, it adds complexity and biases, while newer designs like Mish and Smish showed stability only in specific cases and failed to work in overall cases. 

    To solve these issues, researchers from the University of South Florida proposed a new activation function, TeLU(x) = x · tanh(ex), which combines the learning efficiency of ReLU with the stability and generalization capabilities of smooth functions. This function introduces smooth transitions, which means that the function output changes gradually as the input changes, near-zero-mean activations, and robust gradient dynamics to overcome some of the problems of existing activation functions. The design aims to provide consistent performance across various tasks, improve convergence, and enhance stability with better generalization in shallow and deep architectures.

    Researchers focused on enhancing neural networks while maintaining computational efficiency. Researchers aimed to converge the algorithm quickly, keep it stable during training, and make it robust to generalization over unseen data. The function exists non-polynomially and analytically; hence, it can approximate any continuous target function. The approach emphasized improving learning stability and self-regularization while minimizing numerical instability. By combining linear and non-linear properties, the framework can support efficient learning and help avoid issues such as exploding gradients.

    Researchers evaluated TeLU’s performance through experiments and compared it with other activation functions. The results showed that TeLU helped to prevent the vanishing gradient problem, which is important for effectively training deep networks. It was tested on large datasets such as ImageNet and Dynamic-Pooling Transformers on Text8, showing faster convergence and better accuracy than traditional functions like ReLU. The experiments also showed that TeLU is computationally efficient and works well with ReLU-based configurations, often leading to improved results. The experiments confirmed that TeLU is stable and performs better across various neural network architectures and training methods.

    Hostinger

    In the end, the proposed activation function by the researchers handled key challenges of existing activation functions by preventing the vanishing gradient problem, enhancing computational efficiency, and showing better performance across diverse datasets and architectures. Its successful application on benchmarks like ImageNet, Text8, and Penn Treebank, showing faster convergence, accuracy improvements, and stability in deep learning models, can position TeLU as a promising tool for deep neural networks. Also, TeLU’s performance can serve as a baseline for future research, which can inspire further development of activation functions to achieve even greater efficiency and reliability in machine learning advancements.


    Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

    🚨 FREE UPCOMING AI WEBINAR (JAN 15, 2025): Boost LLM Accuracy with Synthetic Data and Evaluation Intelligence–Join this webinar to gain actionable insights into boosting LLM model performance and accuracy while safeguarding data privacy.

    The post University of South Florida Researchers Propose TeLU Activation Function for Fast and Stable Deep Learning appeared first on MarkTechPost.

    Source: Read More 

    Hostinger
    Facebook Twitter Reddit Email Copy Link
    Previous ArticleKnowledge Sharing Ideas [closed]
    Next Article From Kernels to Attention: Exploring Robust Principal Components in Transformers

    Related Posts

    Artificial Intelligence

    Markus Buehler receives 2025 Washington Award

    June 1, 2025
    Artificial Intelligence

    LWiAI Podcast #201 – GPT 4.5, Sonnet 3.7, Grok 3, Phi 4

    June 1, 2025
    Leave A Reply Cancel Reply

    Continue Reading

    Duo Charged with Operating $430 Million Dark Web Marketplace

    Development

    How to Balance Password Security Against User Experience

    Development

    The best e-commerce website builders of 2025: Expert tested

    News & Updates

    Top 10 Mobile Apps to Boost Your Creativity

    Development

    Highlights

    How to Transfer Your UXR Skills to Different Roles

    March 26, 2025

    Post Content Source: Read More 

    LongPO: Enhancing Long-Context Alignment in LLMs Through Self-Optimized Short-to-Long Preference Learning

    February 26, 2025

    Transcribe audio with Ruby using Universal-1

    April 18, 2024

    Say no to abstract code

    April 24, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.