Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      The Case For Minimal WordPress Setups: A Contrarian View On Theme Frameworks

      June 8, 2025

      How To Fix Largest Contentful Paint Issues With Subpart Analysis

      June 8, 2025

      How To Prevent WordPress SQL Injection Attacks

      June 8, 2025

      AI is currently in its teenage years, battling raging hormones

      June 6, 2025

      The 2025 Wholesome Direct was chock-full of cozy casual games and aesthetic vibes

      June 8, 2025

      All known AMD Ryzen Z2 Series gaming handheld chips and how they are rumored to compare

      June 8, 2025

      “The original Mass Effect trilogy was absolutely an inspiration.” The RPG experts behind WH40K: Rogue Trader are hitting it big with ‘The Expanse: Osiris Reborn’

      June 8, 2025

      If you’ve started to notice Xbox console games showing up on the Xbox PC store, you’re not alone — what’s going on?

      June 8, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Master Image Processing in Node.js Using Sharp for Fast Web Apps

      June 7, 2025
      Recent

      Master Image Processing in Node.js Using Sharp for Fast Web Apps

      June 7, 2025

      mkocansey/bladewind

      June 7, 2025

      Handling PostgreSQL Migrations in Node.js

      June 6, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      The 2025 Wholesome Direct was chock-full of cozy casual games and aesthetic vibes

      June 8, 2025
      Recent

      The 2025 Wholesome Direct was chock-full of cozy casual games and aesthetic vibes

      June 8, 2025

      All known AMD Ryzen Z2 Series gaming handheld chips and how they are rumored to compare

      June 8, 2025

      “The original Mass Effect trilogy was absolutely an inspiration.” The RPG experts behind WH40K: Rogue Trader are hitting it big with ‘The Expanse: Osiris Reborn’

      June 8, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»Carnegie Mellon University at ICLR 2025

    Carnegie Mellon University at ICLR 2025

    April 23, 2025

    CMU researchers are presenting 143 papers at the Thirteenth International Conference on Learning Representations (ICLR 2025), held from April 24 – 28 at the Singapore EXPO. Here is a quick overview of the areas our researchers are working on:

    And here are our most frequent collaborator institutions:

    Table of Contents

    • Oral Papers
    • Spotlight Papers
    • Poster Papers
      • Alignment, Fairness, Safety, Privacy, And Societal Considerations
      • Applications to Computer Vision, Audio, Language, And Other Modalities
      • Applications to Neuroscience & Cognitive Science
      • Applications to Physical Sciences (Physics, Chemistry, Biology, Etc.)
      • Applications to Robotics, Autonomy, Planning
      • Causal Reasoning
      • Datasets and Benchmarks
      • Foundation or Frontier Models, Including LLMs
      • Generative Models
      • Infrastructure, Software Libraries, Hardware, Systems, etc.
      • Interpretability and Explainable AI
      • Learning on Graphs and Other Geometries & Topologies
      • Learning Theory
      • Neurosymbolic & Hybrid AI Systems (Physics-Informed, Logic & Formal Reasoning, etc.)
      • Optimization
      • Other Topics in Machine Learning (i.e., none of the above)
      • Probabilistic Methods (Bayesian Methods, Variational Inference, Sampling, Uncertainty Quantification, etc.)
      • Reinforcement Learning
      • Transfer Learning, Meta Learning, and Lifelong Learning
      • Unsupervised, Self-supervised, Semi-supervised, and Supervised Representation Learning

    Oral Papers

    Backtracking Improves Generation Safety

    Authors: Yiming Zhang, Jianfeng Chi, Hailey Nguyen, Kartikeya Upasani, Daniel M. Bikel, Jason E Weston, Eric Michael Smith

    This paper introduces backtracking, a new technique that allows language models to recover from unsafe text generation by using a special [RESET] token to “undo” problematic outputs. Unlike traditional safety methods that aim to prevent harmful responses outright, backtracking trains the model to self-correct mid-generation. The authors demonstrate that backtracking significantly improves safety without sacrificing helpfulness, and it also provides robustness against several adversarial attacks.

    BigCodeBench: Benchmarking Code Generation with Diverse Function Calls and Complex Instructions

    Authors: Terry Yue Zhuo, Vu Minh Chien, Jenny Chim, Han Hu, Wenhao Yu, Ratnadira Widyasari, Imam Nur Bani Yusuf, Haolan Zhan, Junda He, Indraneil Paul, Simon Brunner, Chen Gong, James Hoang, Armel Randy Zebaze, Xiaoheng Hong, Wen-ding Li, Jean Kaddour, Ming Xu, Zhihan Zhang, Prateek Yadav, Naman Jain, Alex Gu, Zhoujun Cheng, Jiawei Liu, Qian Liu, Zijian Wang, David Lo, Binyuan Hui, Niklas Muennighoff, Daniel Fried, Xiaoning Du, Harm De Vries, Leandro Von Werra

    Recent advances in LLMs have enabled task automation through Python code, but existing benchmarks mainly focus on simple, self-contained tasks. To assess LLMs’ ability to handle more practical challenges requiring diverse and compositional function use, the authors introduce BigCodeBench—a benchmark covering 1,140 tasks across 139 libraries and 7 domains. Each task includes rigorous testing with high branch coverage, and a variant, BigCodeBench-Instruct, reformulates instructions for natural language evaluation. Results from testing 60 LLMs reveal significant performance gaps, highlighting that current models struggle to follow complex instructions and compose function calls accurately compared to human performance.

    Context-Parametric Inversion: Why Instruction Finetuning May Not Actually Improve Context Reliance

    Authors: Sachin Goyal, Christina Baek, J Zico Kolter, Aditi Raghunathan

    LLMs are expected to follow user-provided context, especially when they contain new or conflicting information. While instruction finetuning should improve this ability, the authors uncover a surprising failure mode called context-parametric inversion: models initially rely more on input context, but this reliance decreases as finetuning continues—even as benchmark performance improves. Through controlled experiments and theoretical analysis, the authors trace the cause to training examples where context aligns with pretraining knowledge, reinforcing parametric reliance. They suggest mitigation strategies and highlight this as a key challenge in instruction tuning.

    EmbodiedSAM: Online Segment Any 3D Thing in Real Time

    Authors: Xiuwei Xu, Huangxing Chen, Linqing Zhao, Ziwei Wang, Jie Zhou, Jiwen Lu

    Embodied tasks demand fine-grained 3D perception, which is difficult to achieve due to limited high-quality 3D data. To address this, the authors propose a method that leverages the Segment Anything Model (SAM) for online 3D instance segmentation by transforming 2D masks into 3D-aware queries. Their approach enables real-time object matching across video frames and efficient inference using a similarity matrix. Experiments across multiple datasets show that the method outperforms offline alternatives and generalizes well to new settings with minimal data.

    LLM-SR: Scientific Equation Discovery via Programming with Large Language Models

    Authors: Parshin Shojaee, Kazem Meidani, Shashank Gupta, Amir Barati Farimani, Chandan K. Reddy

    Mathematical equations are remarkably effective at describing natural phenomena, but discovering them from data is challenging due to vast combinatorial search spaces. Existing symbolic regression methods often overlook domain knowledge and rely on limited representations. To address this, the authors propose LLM-SR, a novel approach that uses Large Language Models to generate equation hypotheses informed by scientific priors and refines them through evolutionary search. Evaluated across multiple scientific domains, LLM-SR outperforms existing methods, particularly in generalization, by efficiently exploring the equation space and producing accurate, interpretable models.

    Mind the Gap: Examining the Self-Improvement Capabilities of Large Language Models

    Authors: Yuda Song, Hanlin Zhang, Udaya Ghai, Carson Eisenach, Sham M. Kakade, Dean Foster

    Self-improvement in Large Language Models involves the model verifying its outputs, filtering data accordingly, and using the refined data for further learning. While effective in practice, there has been little theoretical grounding for this technique. This work presents a comprehensive study of LLM self-improvement, introducing a formal framework centered on the generation-verification gap—a key quantity that governs self-improvement. Experiments reveal that this gap scales consistently with pretraining FLOPs across tasks and model families. The authors also explore when and how iterative self-improvement works and offer insights and strategies to enhance it.

    On the Benefits of Memory for Modeling Time-Dependent PDEs

    Authors: Ricardo Buitrago, Tanya Marwah, Albert Gu, Andrej Risteski

    Data-driven methods offer an efficient alternative to traditional numerical solvers for PDEs, but most existing approaches assume Markovian dynamics, limiting their effectiveness when input signals are distorted. Inspired by the Mori-Zwanzig theory, the authors propose MemNO, a Memory Neural Operator that explicitly incorporates past states using structured state-space models and the Fourier Neural Operator. MemNO demonstrates strong performance on various PDE families, especially on low-resolution inputs, achieving over six times lower error than memoryless baselines.

    On the Identification of Temporal Causal Representation with Instantaneous Dependence

    Authors: Zijian Li, Yifan Shen, Kaitao Zheng, Ruichu Cai, Xiangchen Song, Mingming Gong, Guangyi Chen, Kun Zhang

    This work introduces IDOL (Identification framework for Instantaneous Latent dynamics), a method designed to identify latent causal processes in time series data, even when instantaneous relationships are present. Unlike existing methods that require interventions or grouping of observations, IDOL imposes a sparse influence constraint, allowing both time-delayed and instantaneous causal relations to be captured. Through a temporally variational inference architecture and gradient-based sparsity regularization, IDOL effectively estimates latent variables. Experimental results show that IDOL can identify latent causal processes in simulations and real-world human motion forecasting tasks, demonstrating its practical applicability.

    Progressive distillation induces an implicit curriculum

    Authors: Abhishek Panigrahi, Bingbin Liu, Sadhika Malladi, Andrej Risteski, Surbhi Goel

    This work explores the concept of progressive distillation, where a student model learns from intermediate checkpoints of a teacher model, rather than just the final model. The authors identify an “implicit curriculum” that emerges through these intermediate checkpoints, which accelerates the student’s learning and provides a sample complexity benefit. Using sparse parity as a sandbox, they demonstrate that this curriculum imparts valuable learning steps that are unavailable from the final teacher model. The study extends this idea to Transformers trained on probabilistic context-free grammars (PCFGs) and real-world datasets, showing that the teacher progressively teaches the student to capture longer contexts. Both theoretical and empirical results highlight the effectiveness of progressive distillation across different tasks.

    Scaling Laws for Precision

    Authors: Tanishq Kumar, Zachary Ankner, Benjamin Frederick Spector, Blake Bordelon, Niklas Muennighoff, Mansheej Paul, Cengiz Pehlevan, Christopher Re, Aditi Raghunathan

    This work introduces precision-aware scaling laws that extend traditional scaling frameworks to account for the effects of low-precision training and inference in language models. The authors show that lower precision effectively reduces a model’s usable parameter count, enabling predictions of performance degradation due to quantization. For inference, they find that post-training quantization causes increasing degradation with more pretraining data, potentially making additional training counterproductive. Their unified framework predicts loss across varying precisions and suggests that training larger models in lower precision may be more compute-efficient. These predictions are validated on over 465 pretraining runs, including models up to 1.7B parameters.

    Self-Improvement in Language Models: The Sharpening Mechanism

    Authors: Audrey Huang, Adam Block, Dylan J Foster, Dhruv Rohatgi, Cyril Zhang, Max Simchowitz, Jordan T. Ash, Akshay Krishnamurthy

    This paper presents a theoretical framework for understanding how LLMs can self-improve by using themselves as verifiers to refine their own outputs; a process the authors call “sharpening.” The key insight is that LLMs are often better at judging response quality than generating high-quality responses outright, so sharpening helps concentrate probability mass on better sequences. The paper analyzes two families of self-improvement algorithms: one based on supervised fine-tuning (SFT) and one on reinforcement learning (RLHF). They show that while the SFT-based approach is optimal under certain conditions, the RLHF-based approach can outperform it by actively exploring beyond the model’s existing knowledge.

    When Selection meets Intervention: Additional Complexities in Causal Discovery

    Authors: Haoyue Dai, Ignavier Ng, Jianle Sun, Zeyu Tang, Gongxu Luo, Xinshuai Dong, Peter Spirtes, Kun Zhang

    This work tackles the often-overlooked issue of selection bias in interventional studies, where participants are selectively included based on specific criteria. Existing causal discovery methods typically ignore this bias, leading to inaccurate conclusions. To address this, the authors introduce a novel graphical model that distinguishes between the observed world with interventions and the counterfactual world where selection occurs. They develop a sound algorithm that identifies both causal relationships and selection mechanisms, demonstrating its effectiveness through experiments on both synthetic and real-world data.

    miniCTX: Neural Theorem Proving with (Long-)Contexts

    Authors: Jiewen Hu, Thomas Zhu, Sean Welleck

    Real-world formal theorem proving relies heavily on rich contextual information, which is often absent from traditional benchmarks. To address this, the authors introduce miniCTX, a benchmark designed to test models’ ability to prove theorems using previously unseen, extensive context from real Lean projects and textbooks. Unlike prior benchmarks, miniCTX includes large repositories with relevant definitions, lemmas, and structures. Baseline experiments show that models conditioned on this broader context significantly outperform those relying solely on the local state. The authors also provide a toolkit to facilitate the expansion of the benchmark.

    Spotlight Papers

    ADIFF: Explaining audio difference using natural language

    Authors: Soham Deshmukh, Shuo Han, Rita Singh, Bhiksha Raj

    This paper tackles the novel task of explaining differences between audio recordings, which is important for applications like audio forensics, quality assessment, and generative audio systems. The authors introduce two new datasets and propose a three-tiered explanation framework—ranging from concise event descriptions to rich, emotionally grounded narratives—generated using large language models. They present ADIFF, a new method that improves on baselines by incorporating audio cross-projection, position-aware captioning, and multi-stage training, and show that it significantly outperforms existing audio-language models both quantitatively and via human evaluation.

    Better Instruction-Following Through Minimum Bayes Risk

    Authors: Ian Wu, Patrick Fernandes, Amanda Bertsch, Seungone Kim, Sina Khoshfetrat Pakazad, Graham Neubig

    This paper explores how LLMs can be used as judges to evaluate and improve other LLMs. The authors show that using a method called Minimum Bayes Risk (MBR) decoding—where an LLM judge selects the best output from a set—can significantly improve model performance compared to standard decoding methods. They also find that training models on these high-quality outputs can lead to strong gains even without relying on MBR at test time, making the models faster and more efficient while maintaining or exceeding previous performance.

    DeFT: Decoding with Flash Tree-attention for Efficient Tree-structured LLM Inference

    Authors: Jinwei Yao, Kaiqi Chen, Kexun Zhang, Jiaxuan You, Binhang Yuan, Zeke Wang, Tao Lin

    This paper introduces DeFT, a new algorithm that speeds up how large language models handle tasks involving tree-like structures with shared text prefixes, such as multi-step reasoning or few-shot prompting. Existing methods waste time and memory by repeatedly accessing the same data and poorly distributing the workload across the GPU. DeFT solves this by smartly grouping and splitting memory usage to avoid redundant operations and better balance the work, leading to up to 3.6x faster performance on key tasks compared to current approaches.

    Holistically Evaluating the Environmental Impact of Creating Language Models

    Authors: Jacob Morrison, Clara Na, Jared Fernandez, Tim Dettmers, Emma Strubell, Jesse Dodge

    This paper estimates the full environmental impact of developing large language models, including not just the final training runs but also model development and hardware manufacturing—areas typically underreported. The authors found that training a series of models released 493 metric tons of carbon emissions and used 2.769 million liters of water, even in a highly efficient data center. Notably, around half of the carbon emissions came from the development phase alone, and power usage during training varied significantly, raising concerns for energy grid planning as AI systems grow.

    Language Model Alignment in Multilingual Trolley Problems

    Authors: Zhijing Jin, Max Kleiman-weiner, Giorgio Piatti, Sydney Levine, Jiarui Liu, Fernando Gonzalez Adauto, Francesco Ortu, András Strausz, Mrinmaya Sachan, Rada Mihalcea, Yejin Choi, Bernhard Schölkopf

    This paper evaluates how well LLMs align with human moral preferences across languages using multilingual trolley problems. The authors introduce MultiTP, a new dataset of moral dilemmas in over 100 languages based on the Moral Machine experiment, enabling cross-lingual analysis of LLM decision-making. By assessing 19 models across six moral dimensions and examining demographic correlations and prompt consistency, they uncover significant variation in moral alignment across languages—highlighting ethical biases and the need for more inclusive, multilingual approaches to responsible AI development.

    Lean-STaR: Learning to Interleave Thinking and Proving

    Authors: Haohan Lin, Zhiqing Sun, Sean Welleck, Yiming Yang

    This paper introduces Lean-STaR, a framework that improves language model-based theorem proving by incorporating informal “thoughts” before each proof step. Unlike traditional approaches that rely solely on formal proof data, Lean-STaR generates synthetic thought processes using retrospective proof tactics during training. At inference time, the model generates these thoughts to guide its next action, and expert iteration further refines its performance using the Lean theorem prover. This approach boosts proof success rates and offers new insights into how structured reasoning improves formal mathematical problem solving.

    MagicPIG: LSH Sampling for Efficient LLM Generation

    Authors: Zhuoming Chen, Ranajoy Sadhukhan, Zihao Ye, Yang Zhou, Jianyu Zhang, Niklas Nolte, Yuandong Tian, Matthijs Douze, Leon Bottou, Zhihao Jia, Beidi Chen

    This paper introduces MagicPIG, a new system that speeds up LLM inference by approximating attention more efficiently. While many methods assume attention is sparse and use TopK approximations, the authors show this isn’t always accurate and can hurt performance. Instead, MagicPIG uses a sampling method backed by theoretical guarantees and accelerates it using Locality Sensitive Hashing, offloading computations to the CPU to support longer inputs and larger batches without sacrificing accuracy.

    Multi-Robot Motion Planning with Diffusion Models

    Authors: Yorai Shaoul, Itamar Mishani, Shivam Vats, Jiaoyang Li, Maxim Likhachev

    This paper introduces a method for planning coordinated, collision-free movements for many robots using only data from individual robots. The authors combine learned diffusion models with classical planning algorithms to generate realistic, safe multi-robot trajectories. Their approach, called Multi-robot Multi-model planning Diffusion, also scales to large environments by stitching together multiple diffusion models, showing strong results in simulated logistics scenarios.

    Reinforcement Learning for Control of Non-Markovian Cellular Population Dynamics

    Authors: Josiah C Kratz, Jacob Adamczyk

    This paper explores how reinforcement learning can be used to develop drug dosing strategies for controlling cell populations that adapt over time, such as cancer cells switching between resistant and susceptible states. Traditional methods struggle when the system’s dynamics are unknown or involve memory of past environments, making optimal control difficult. The authors show that deep RL can successfully learn effective strategies even in complex, memory-based systems, offering a promising approach for real-world biomedical applications.

    Rewarding Progress: Scaling Automated Process Verifiers for LLM Reasoning

    Authors: Amrith Setlur, Chirag Nagpal, Adam Fisch, Xinyang Geng, Jacob Eisenstein, Rishabh Agarwal, Alekh Agarwal, Jonathan Berant, Aviral Kumar

    This paper explores how to improve large language models’ reasoning by giving feedback at each step of their thinking process, rather than only at the final answer. The authors introduce a method where feedback—called a process reward—is based on whether a step helps make a correct final answer more likely, as judged by a separate model (a “prover”) that can recognize progress better than the model being trained. They show both theoretically and experimentally that this strategy makes learning more efficient, leading to significantly better and faster results than traditional outcome-based feedback methods.

    SVDQuant: Absorbing Outliers by Low-Rank Component for 4-Bit Diffusion Models

    Authors: Muyang Li, Yujun Lin, Zhekai Zhang, Tianle Cai, Junxian Guo, Xiuyu Li, Enze Xie, Chenlin Meng, Jun-yan Zhu, Song Han

    This paper introduces SVDQuant, a method for significantly speeding up diffusion models by quantizing both weights and activations to 4 bits. Since such aggressive quantization can hurt image quality, the authors use a clever technique: they shift problematic “outlier” values into a separate low-rank component handled with higher precision, while the rest is processed with efficient low-bit operations. To avoid slowing things down due to extra computation, they also design a custom inference engine called Nunchaku, which merges the processing steps to minimize memory access. Together, these techniques reduce memory usage and deliver over 3x speedups without sacrificing image quality.

    Stabilizing Reinforcement Learning in Differentiable Multiphysics Simulation

    Authors: Eliot Xing, Vernon Luk, Jean Oh

    This paper tackles the challenge of applying reinforcement learning (RL) to soft-body robotics, where simulations are usually too slow for data-hungry RL algorithms. The authors introduce SAPO, a new model-based RL algorithm that efficiently learns from differentiable simulations using analytic gradients. The authors also present Rewarped, a fast, parallel simulation platform that supports both rigid and deformable materials, demonstrating that their approach outperforms existing methods on complex manipulation and locomotion tasks.

    Streaming Algorithms For $ell_p$ Flows and $ell_p$ Regression

    Authors: Amit Chakrabarti, Jeffrey Jiang, David Woodruff, Taisuke Yasuda

    This paper investigates how to solve underdetermined linear regression problems in a streaming setting, where the data arrives one column at a time and storing the full dataset is impractical. The authors develop algorithms that approximate the regression cost or output a near-optimal solution using much less memory than storing the entire dataset—particularly relevant for applications like computing flows on large graphs. They also establish space lower bounds, showing the limitations of what’s possible, and provide the first algorithms that achieve nontrivial approximations using sublinear space in various settings.

    Poster Papers

    Alignment, Fairness, Safety, Privacy, And Societal Considerations

    $beta$-calibration of Language Model Confidence Scores for Generative QA

    Authors: Putra Manggala, Atalanti A. Mastakouri, Elke Kirschbaum, Shiva Kasiviswanathan, Aaditya Ramdas

    AgentHarm: Benchmarking Robustness of LLM Agents on Harmful Tasks

    Authors: Maksym Andriushchenko, Alexandra Souly, Mateusz Dziemian, Derek Duenas, Maxwell Lin, Justin Wang, Dan Hendrycks, Andy Zou, J Zico Kolter, Matt Fredrikson, Yarin Gal, Xander Davies

    Aligned LLMs Are Not Aligned Browser Agents

    Authors: Priyanshu Kumar, Elaine Lau, Saranya Vijayakumar, Tu Trinh, Elaine T Chang, Vaughn Robinson, Shuyan Zhou, Matt Fredrikson, Sean M. Hendryx, Summer Yue, Zifan Wang

    Comparing Targeting Strategies for Maximizing Social Welfare with Limited Resources

    Authors: Vibhhu Sharma, Bryan Wilder

    Concept-ROT: Poisoning Concepts in Large Language Models with Model Editing

    Authors: Keltin Grimes, Marco Christiani, David Shriver, Marissa Catherine Connor

    DECISION-FOCUSED UNCERTAINTY QUANTIFICATION

    Authors: Santiago Cortes-gomez, Carlos Miguel Patiño, Yewon Byun, Steven Wu, Eric Horvitz, Bryan Wilder

    Dissecting Adversarial Robustness of Multimodal LM Agents

    Authors: Chen Henry Wu, Rishi Rajesh Shah, Jing Yu Koh, Russ Salakhutdinov, Daniel Fried, Aditi Raghunathan

    Follow My Instruction and Spill the Beans: Scalable Data Extraction from Retrieval-Augmented Generation Systems

    Authors: Zhenting Qi, Hanlin Zhang, Eric P. Xing, Sham M. Kakade, Himabindu Lakkaraju

    Generative Classifiers Avoid Shortcut Solutions

    Authors: Alexander Cong Li, Ananya Kumar, Deepak Pathak

    Jogging the Memory of Unlearned LLMs Through Targeted Relearning Attacks

    Authors: Shengyuan Hu, Yiwei Fu, Steven Wu, Virginia Smith

    Noisy Test-Time Adaptation in Vision-Language Models

    Authors: Chentao Cao, Zhun Zhong, Zhanke Zhou, Tongliang Liu, Yang Liu, Kun Zhang, Bo Han

    Pacmann: Efficient Private Approximate Nearest Neighbor Search

    Authors: Mingxun Zhou, Elaine Shi, Giulia Fanti

    Permute-and-Flip: An optimally stable and watermarkable decoder for LLMs

    Authors: Xuandong Zhao, Lei Li, Yu-xiang Wang

    Persistent Pre-training Poisoning of LLMs

    Authors: Yiming Zhang, Javier Rando, Ivan Evtimov, Jianfeng Chi, Eric Michael Smith, Nicholas Carlini, Florian Tramèr, Daphne Ippolito

    Prompting Fairness: Integrating Causality to Debias Large Language Models

    Authors: Jingling Li, Zeyu Tang, Xiaoyu Liu, Peter Spirtes, Kun Zhang, Liu Leqi, Yang Liu

    Reconciling Model Multiplicity for Downstream Decision Making

    Authors: Ally Yalei Du, Dung Daniel Ngo, Steven Wu

    Self-Play Preference Optimization for Language Model Alignment

    Authors: Yue Wu, Zhiqing Sun, Huizhuo Yuan, Kaixuan Ji, Yiming Yang, Quanquan Gu

    Toward Robust Defenses Against LLM Weight Tampering Attacks

    Authors: Rishub Tamirisa, Bhrugu Bharathi, Long Phan, Andy Zhou, Alice Gatti, Tarun Suresh, Maxwell Lin, Justin Wang, Rowan Wang, Ron Arel, Andy Zou, Dawn Song, Bo Li, Dan Hendrycks, Mantas Mazeika

    Applications To Computer Vision, Audio, Language, And Other Modalities

    Agent-to-Sim: Learning Interactive Behavior Model from Casual Longitudinal Videos

    Authors: Gengshan Yang, Andrea Bajcsy, Shunsuke Saito, Angjoo Kanazawa

    Context-aware Dynamic Pruning for Speech Foundation Models

    Authors: Masao Someki, Yifan Peng, Siddhant Arora, Markus Müller, Athanasios Mouchtaris, Grant Strimel, Jing Liu, Shinji Watanabe

    Diff-2-in-1: Bridging Generation and Dense Perception with Diffusion Models

    Authors: Shuhong Zheng, Zhipeng Bao, Ruoyu Zhao, Martial Hebert, Yu-xiong Wang

    Fugatto 1: Foundational Generative Audio Transformer Opus 1

    Authors: Rafael Valle, Rohan Badlani, Zhifeng Kong, Sang-gil Lee, Arushi Goel, Joao Felipe Santos, Aya Aljafari, Sungwon Kim, Shuqi Dai, Siddharth Gururani, Alexander H. Liu, Kevin J. Shih, Ryan Prenger, Wei Ping, Chao-han Huck Yang, Bryan Catanzaro

    Gaussian Splatting Lucas-Kanade

    Authors: Liuyue Xie, Joel Julin, Koichiro Niinuma, Laszlo Attila Jeni

    ImageFolder: Autoregressive Image Generation with Folded Tokens

    Authors: Xiang Li, Kai Qiu, Hao Chen, Jason Kuen, Jiuxiang Gu, Bhiksha Raj, Zhe Lin

    Improving Large Language Model based Multi-Agent Framework through Dynamic Workflow Updating

    Authors: Boye Niu, Yiliao Song, Kai Lian, Yifan Shen, Yu Yao, Kun Zhang, Tongliang Liu

    MetaDesigner: Advancing Artistic Typography through AI-Driven, User-Centric, and Multilingual WordArt Synthesis

    Authors: Jun-yan He, Zhi-qi Cheng, Chenyang Li, Jingdong Sun, Qi He, Wangmeng Xiang, Hanyuan Chen, Jin-peng Lan, Xianhui Lin, Kang Zhu, Bin Luo, Yifeng Geng, Xuansong Xie, Alexander G Hauptmann

    OMG: Opacity Matters in Material Modeling with Gaussian Splatting

    Authors: Silong Yong, Venkata Nagarjun Pudureddiyur Manivannan, Bernhard Kerbl, Zifu Wan, Simon Stepputtis, Katia P. Sycara, Yaqi Xie

    Scene Flow as a Partial Differential Equation

    Authors: Kyle Vedder, Neehar Peri, Ishan Khatri, Siyi Li, Eric Eaton, Mehmet Kemal Kocamaz, Yue Wang, Zhiding Yu, Deva Ramanan, Joachim Pehserl

    TrackTheMind: program-guided adversarial data generation for theory of mind reasoning

    Authors: Melanie Sclar, Jane Dwivedi-yu, Maryam Fazel-zarandi, Yulia Tsvetkov, Yonatan Bisk, Yejin Choi, Asli Celikyilmaz

    Understanding Visual Concepts Across Models

    Authors: Brandon Trabucco, Max A Gurinas, Kyle Doherty, Russ Salakhutdinov

    Applications To Neuroscience & Cognitive Science

    Brain Mapping with Dense Features: Grounding Cortical Semantic Selectivity in Natural Images With Vision Transformers

    Authors: Andrew Luo, Jacob Yeung, Rushikesh Zawar, Shaurya Rajat Dewan, Margaret Marie Henderson, Leila Wehbe, Michael J. Tarr

    Self-Attention-Based Contextual Modulation Improves Neural System Identification

    Authors: Isaac Lin, Tianye Wang, Shang Gao, Tang Shiming, Tai Sing Lee

    Applications To Physical Sciences (Physics, Chemistry, Biology, Etc.)

    Causal Representation Learning from Multimodal Biological Observations

    Authors: Yuewen Sun, Lingjing Kong, Guangyi Chen, Loka Li, Gongxu Luo, Zijian Li, Yixuan Zhang, Yujia Zheng, Mengyue Yang, Petar Stojanov, Eran Segal, Eric P. Xing, Kun Zhang

    Chemistry-Inspired Diffusion with Non-Differentiable Guidance

    Authors: Yuchen Shen, Chenhao Zhang, Sijie Fu, Chenghui Zhou, Newell Washburn, Barnabas Poczos

    Text2PDE: Latent Diffusion Models for Accessible Physics Simulation

    Authors: Anthony Zhou, Zijie Li, Michael Schneier, John R Buchanan Jr, Amir Barati Farimani

    Applications To Robotics, Autonomy, Planning

    Enhancing Software Agents with Monte Carlo Tree Search and Hindsight Feedback

    Authors: Antonis Antoniades, Albert Örwall, Kexun Zhang, Yuxi Xie, Anirudh Goyal, William Yang Wang

    ThinkBot: Embodied Instruction Following with Thought Chain Reasoning

    Authors: Guanxing Lu, Ziwei Wang, Changliu Liu, Jiwen Lu, Yansong Tang

    Causal Reasoning

    A Conditional Independence Test in the Presence of Discretization

    Authors: Boyang Sun, Yu Yao, Guang-yuan Hao, Yumou Qiu, Kun Zhang

    A Robust Method to Discover Causal or Anticausal Relation

    Authors: Yu Yao, Yang Zhou, Bo Han, Mingming Gong, Kun Zhang, Tongliang Liu

    A Skewness-Based Criterion for Addressing Heteroscedastic Noise in Causal Discovery

    Authors: Yingyu Lin, Yuxing Huang, Wenqin Liu, Haoran Deng, Ignavier Ng, Kun Zhang, Mingming Gong, Yian Ma, Biwei Huang

    Analytic DAG Constraints for Differentiable DAG Learning

    Authors: Zhen Zhang, Ignavier Ng, Dong Gong, Yuhang Liu, Mingming Gong, Biwei Huang, Kun Zhang, Anton Van Den Hengel, Javen Qinfeng Shi

    Causal Graph Transformer for Treatment Effect Estimation Under Unknown Interference

    Authors: Anpeng Wu, Haiyi Qiu, Zhengming Chen, Zijian Li, Ruoxuan Xiong, Fei Wu, Kun Zhang

    Differentiable Causal Discovery for Latent Hierarchical Causal Models

    Authors: Parjanya Prajakta Prashant, Ignavier Ng, Kun Zhang, Biwei Huang

    Datasets And Benchmarks

    Dynamic-SUPERB Phase-2: A Collaboratively Expanding Benchmark for Measuring the Capabilities of Spoken Language Models with 180 Tasks

    Authors: Chien-yu Huang, Wei-chih Chen, Shu-wen Yang, Andy T. Liu, Chen-an Li, Yu-xiang Lin, Wei-cheng Tseng, Anuj Diwan, Yi-jen Shih, Jiatong Shi, William Chen, Xuanjun Chen, Chi-yuan Hsiao, Puyuan Peng, Shih-heng Wang, Chun-yi Kuan, Ke-han Lu, Kai-wei Chang, Chih-kai Yang, Fabian Alejandro Ritter Gutierrez, Huang Kuan-po, Siddhant Arora, You-kuan Lin, Chuang Ming To, Eunjung Yeo, Kalvin Chang, Chung-ming Chien, Kwanghee Choi, Cheng-hsiu Hsieh, Yi-cheng Lin, Chee-en Yu, I-hsiang Chiu, Heitor Guimarães, Jionghao Han, Tzu-quan Lin, Tzu-yuan Lin, Homu Chang, Ting-wu Chang, Chun Wei Chen, Shou-jen Chen, Yu-hua Chen, Hsi-chun Cheng, Kunal Dhawan, Jia-lin Fang, Shi-xin Fang, Kuan Yu Fang Chiang, Chi An Fu, Hsien-fu Hsiao, Ching Yu Hsu, Shao-syuan Huang, Lee Chen Wei, Hsi-che Lin, Hsuan-hao Lin, Hsuan-ting Lin, Jian-ren Lin, Ting-chun Liu, Li-chun Lu, Tsung-min Pai, Ankita Pasad, Shih-yun Shan Kuan, Suwon Shon, Yuxun Tang, Yun-shao Tsai, Wei Jui Chiang, Tzu-chieh Wei, Chengxi Wu, Dien-ruei Wu, Chao-han Huck Yang, Chieh-chi Yang, Jia Qi Yip, Shao-xiang Yuan, Haibin Wu, Karen Livescu, David Harwath, Shinji Watanabe, Hung-yi Lee

    GameArena: Evaluating LLM Reasoning through Live Computer Games

    Authors: Lanxiang Hu, Qiyu Li, Anze Xie, Nan Jiang, Ion Stoica, Haojian Jin, Hao Zhang

    Harnessing Webpage UIs for Text-Rich Visual Understanding

    Authors: Junpeng Liu, Tianyue Ou, Yifan Song, Yuxiao Qu, Wai Lam, Chenyan Xiong, Wenhu Chen, Graham Neubig, Xiang Yue

    Scalable Benchmarking and Robust Learning for Noise-Free Ego-Motion and 3D Reconstruction from Noisy Video

    Authors: Xiaohao Xu, Tianyi Zhang, Shibo Zhao, Xiang Li, Sibo Wang, Yongqi Chen, Ye Li, Bhiksha Raj, Matthew Johnson-roberson, Sebastian Scherer, Xiaonan Huang

    Speech Robust Bench: A Robustness Benchmark For Speech Recognition

    Authors: Muhammad A Shah, David Solans Noguero, Mikko A. Heikkilä, Bhiksha Raj, Nicolas Kourtellis

    Talking Turns: Benchmarking Audio Foundation Models on Turn-Taking Dynamics

    Authors: Siddhant Arora, Zhiyun Lu, Chung-cheng Chiu, Ruoming Pang, Shinji Watanabe

    TestGenEval: A Real World Unit Test Generation and Test Completion Benchmark

    Authors: Kush Jain, Gabriel Synnaeve, Baptiste Roziere

    Unearthing Skill-level Insights for Understanding Trade-offs of Foundation Models

    Authors: Mazda Moayeri, Vidhisha Balachandran, Varun Chandrasekaran, Safoora Yousefi, Thomas Fel, Soheil Feizi, Besmira Nushi, Neel Joshi, Vibhav Vineet

    VideoWebArena: Evaluating Long Context Multimodal Agents with Video Understanding Web Tasks

    Authors: Lawrence Keunho Jang, Yinheng Li, Dan Zhao, Charles Ding, Justin Lin, Paul Pu Liang, Rogerio Bonatti, Kazuhito Koishida

    Foundation Or Frontier Models, Including Llms

    Adaptive Data Optimization: Dynamic Sample Selection with Scaling Laws

    Authors: Yiding Jiang, Allan Zhou, Zhili Feng, Sadhika Malladi, J Zico Kolter

    Diversity Empowers Intelligence: Integrating Expertise of Software Engineering Agents

    Authors: Kexun Zhang, Weiran Yao, Zuxin Liu, Yihao Feng, Zhiwei Liu, Rithesh R N, Tian Lan, Lei Li, Renze Lou, Jiacheng Xu, Bo Pang, Yingbo Zhou, Shelby Heinecke, Silvio Savarese, Huan Wang, Caiming Xiong

    Generalization v.s. Memorization: Tracing Language Models’ Capabilities Back to Pretraining Data

    Authors: Xinyi Wang, Antonis Antoniades, Yanai Elazar, Alfonso Amayuelas, Alon Albalak, Kexun Zhang, William Yang Wang

    Improving Large Language Model Planning with Action Sequence Similarity

    Authors: Xinran Zhao, Hanie Sedghi, Bernd Bohnet, Dale Schuurmans, Azade Nova

    Inference Optimal VLMs Need Only One Visual Token but Larger Models

    Authors: Kevin Li, Sachin Goyal, João D. Semedo, J Zico Kolter

    Inference Scaling Laws: An Empirical Analysis of Compute-Optimal Inference for LLM Problem-Solving

    Authors: Yangzhen Wu, Zhiqing Sun, Shanda Li, Sean Welleck, Yiming Yang

    Language Models Need Inductive Biases to Count Inductively

    Authors: Yingshan Chang, Yonatan Bisk

    MIND: Math Informed syNthetic Dialogues for Pretraining LLMs

    Authors: Syeda Nahida Akter, Shrimai Prabhumoye, John Kamalu, Sanjeev Satheesh, Eric Nyberg, Mostofa Patwary, Mohammad Shoeybi, Bryan Catanzaro

    MatryoshkaKV: Adaptive KV Compression via Trainable Orthogonal Projection

    Authors: Bokai Lin, Zihao Zeng, Zipeng Xiao, Siqi Kou, Tianqi Hou, Xiaofeng Gao, Hao Zhang, Zhijie Deng

    Mixture of Parrots: Experts improve memorization more than reasoning

    Authors: Samy Jelassi, Clara Mohri, David Brandfonbrener, Alex Gu, Nikhil Vyas, Nikhil Anand, David Alvarez-melis, Yuanzhi Li, Sham M. Kakade, Eran Malach

    Montessori-Instruct: Generate Influential Training Data Tailored for Student Learning

    Authors: Xiaochuan Li, Zichun Yu, Chenyan Xiong

    Pangea: A Fully Open Multilingual Multimodal LLM for 39 Languages

    Authors: Xiang Yue, Yueqi Song, Akari Asai, Simran Khanuja, Anjali Kantharuban, Seungone Kim, Jean De Dieu Nyandwi, Lintang Sutawika, Sathyanarayanan Ramamoorthy, Graham Neubig

    Physics of Language Models: Part 3.2, Knowledge Manipulation

    Authors: Zeyuan Allen-zhu, Yuanzhi Li

    Scaling Long Context Training Data by Long-Distance Referrals

    Authors: Yonghao Zhuang, Lanxiang Hu, Longfei Yun, Souvik Kundu, Zhengzhong Liu, Eric P. Xing, Hao Zhang

    Sparse Matrix in Large Language Model Fine-tuning

    Authors: Haoze He, Juncheng B Li, Xuan Jiang, Heather Miller

    Specialized Foundation Models struggle to beat Supervised Baselines

    Authors: Zongzhe Xu, Ritvik Gupta, Wenduo Cheng, Alexander Shen, Junhong Shen, Ameet Talwalkar, Mikhail Khodak

    Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo

    Authors: Shengyu Feng, Xiang Kong, Shuang Ma, Aonan Zhang, Dong Yin, Chong Wang, Ruoming Pang, Yiming Yang

    TidalDecode: Fast and Accurate LLM Decoding with Position Persistent Sparse Attention

    Authors: Lijie Yang, Zhihao Zhang, Zhuofu Chen, Zikun Li, Zhihao Jia

    Time, Space and Streaming Efficient Algorithm for Heavy Attentions

    Authors: Ravindran Kannan, Chiranjib Bhattacharyya, Praneeth Kacham, David Woodruff

    Generative Models

    APE: Faster and Longer Context-Augmented Generation via Adaptive Parallel Encoding

    Authors: Xinyu Yang, Tianqi Chen, Beidi Chen

    Consistency Models Made Easy

    Authors: Zhengyang Geng, Ashwini Pokle, Weijian Luo, Justin Lin, J Zico Kolter

    Human-Aligned Chess With a Bit of Search

    Authors: Yiming Zhang, Athul Paul Jacob, Vivian Lai, Daniel Fried, Daphne Ippolito

    Linear Combination of Saved Checkpoints Makes Consistency and Diffusion Models Better

    Authors: Enshu Liu, Junyi Zhu, Zinan Lin, Xuefei Ning, Shuaiqi Wang, Matthew B. Blaschko, Sergey Yekhanin, Shengen Yan, Guohao Dai, Huazhong Yang, Yu Wang

    MagicDec: Breaking the Latency-Throughput Tradeoff for Long Context Generation with Speculative Decoding

    Authors: Ranajoy Sadhukhan, Jian Chen, Zhuoming Chen, Vashisth Tiwari, Ruihang Lai, Jinyuan Shi, Ian En-hsu Yen, Avner May, Tianqi Chen, Beidi Chen

    OmniPhysGS: 3D Constitutive Gaussians for General Physics-Based Dynamics Generation

    Authors: Yuchen Lin, Chenguo Lin, Jianjin Xu, Yadong Mu

    RAG-DDR: Optimizing Retrieval-Augmented Generation Using Differentiable Data Rewards

    Authors: Xinze Li, Sen Mei, Zhenghao Liu, Yukun Yan, Shuo Wang, Shi Yu, Zheni Zeng, Hao Chen, Ge Yu, Zhiyuan Liu, Maosong Sun, Chenyan Xiong

    Speculative Knowledge Distillation: Bridging the Teacher-Student Gap Through Interleaved Sampling

    Authors: Wenda Xu, Rujun Han, Zifeng Wang, Long Le, Dhruv Madeka, Lei Li, William Yang Wang, Rishabh Agarwal, Chen-yu Lee, Tomas Pfister

    TFG-Flow: Training-free Guidance in Multimodal Generative Flow

    Authors: Haowei Lin, Shanda Li, Haotian Ye, Yiming Yang, Stefano Ermon, Yitao Liang, Jianzhu Ma

    Truncated Consistency Models

    Authors: Sangyun Lee, Yilun Xu, Tomas Geffner, Giulia Fanti, Karsten Kreis, Arash Vahdat, Weili Nie

    TypedThinker: Typed Thinking Improves Large Language Model Reasoning

    Authors: Danqing Wang, Jianxin Ma, Fei Fang, Lei Li

    Infrastructure, Software Libraries, Hardware, Systems, Etc.

    OpenHands: An Open Platform for AI Software Developers as Generalist Agents

    Authors: Xingyao Wang, Boxuan Li, Yufan Song, Frank F. Xu, Xiangru Tang, Mingchen Zhuge, Jiayi Pan, Yueqi Song, Bowen Li, Jaskirat Singh, Hoang H. Tran, Fuqiang Li, Ren Ma, Mingzhang Zheng, Bill Qian, Yanjun Shao, Niklas Muennighoff, Yizhe Zhang, Binyuan Hui, Junyang Lin, Robert Brennan, Hao Peng, Heng Ji, Graham Neubig

    Interpretability And Explainable Ai

    Improving Instruction-Following in Language Models through Activation Steering

    Authors: Alessandro Stolfo, Vidhisha Balachandran, Safoora Yousefi, Eric Horvitz, Besmira Nushi

    Interpreting Language Reward Models via Contrastive Explanations

    Authors: Junqi Jiang, Tom Bewley, Saumitra Mishra, Freddy Lecue, Manuela Veloso

    LICORICE: Label-Efficient Concept-Based Interpretable Reinforcement Learning

    Authors: Zhuorui Ye, Stephanie Milani, Geoffrey J. Gordon, Fei Fang

    Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process

    Authors: Tian Ye, Zicheng Xu, Yuanzhi Li, Zeyuan Allen-zhu

    Physics of Language Models: Part 2.2, How to Learn From Mistakes on Grade-School Math Problems

    Authors: Tian Ye, Zicheng Xu, Yuanzhi Li, Zeyuan Allen-zhu

    Sparse autoencoders reveal selective remapping of visual concepts during adaptation

    Authors: Hyesu Lim, Jinho Choi, Jaegul Choo, Steffen Schneider

    Learning On Graphs And Other Geometries & Topologies

    Learning Graph Invariance by Harnessing Spuriosity

    Authors: Tianjun Yao, Yongqiang Chen, Kai Hu, Tongliang Liu, Kun Zhang, Zhiqiang Shen

    Spectro-Riemannian Graph Neural Networks

    Authors: Karish Grover, Haiyang Yu, Xiang Song, Qi Zhu, Han Xie, Vassilis N. Ioannidis, Christos Faloutsos

    Learning Theory

    A Theoretical Analysis of Self-Supervised Learning for Vision Transformers

    Authors: Yu Huang, Zixin Wen, Yuejie Chi, Yingbin Liang

    Adversarial Training Can Provably Improve Robustness: Theoretical Analysis of Feature Learning Process Under Structured Data

    Authors: Binghui Li, Yuanzhi Li

    Larger Language Models Provably Generalize Better

    Authors: Marc Anton Finzi, Sanyam Kapoor, Diego Granziol, Anming Gu, Andrew Gordon Wilson, Christopher De Sa, J Zico Kolter

    Learning from weak labelers as constraints

    Authors: Vishwajeet Agrawal, Rattana Pukdee, Maria Florina Balcan, Pradeep Kumar Ravikumar

    Misspecified $Q$-Learning with Sparse Linear Function Approximation: Tight Bounds on Approximation Error

    Authors: Ally Yalei Du, Lin Yang, Ruosong Wang

    Neurosymbolic & Hybrid Ai Systems (Physics-informed, Logic & Formal Reasoning, Etc.)

    ImProver: Agent-Based Automated Proof Optimization

    Authors: Riyaz Ahuja, Jeremy Avigad, Prasad Tetali, Sean Welleck

    NeSyC: A Neuro-symbolic Continual Learner For Complex Embodied Tasks in Open Domains

    Authors: Wonje Choi, Jinwoo Park, Sanghyun Ahn, Daehee Lee, Honguk Woo

    Optimization

    Understanding Optimization in Deep Learning with Central Flows

    Authors: Jeremy Cohen, Alex Damian, Ameet Talwalkar, J Zico Kolter, Jason D. Lee

    Other Topics In Machine Learning (I.e., None Of The Above)

    AnoLLM: Large Language Models for Tabular Anomaly Detection

    Authors: Che-ping Tsai, Ganyu Teng, Phillip Wallis, Wei Ding

    Beyond Worst-Case Dimensionality Reduction for Sparse Vectors

    Authors: Sandeep Silwal, David Woodruff, Qiuyi Zhang

    Zeroth-Order Fine-Tuning of LLMs with Transferable Static Sparsity

    Authors: Wentao Guo, Jikai Long, Yimeng Zeng, Zirui Liu, Xinyu Yang, Yide Ran, Jacob R. Gardner, Osbert Bastani, Christopher De Sa, Xiaodong Yu, Beidi Chen, Zhaozhuo Xu

    Probabilistic Methods (Bayesian Methods, Variational Inference, Sampling, Uq, Etc.)

    Conformalized Interactive Imitation Learning: Handling Expert Shift and Intermittent Feedback

    Authors: Michelle D Zhao, Henny Admoni, Reid Simmons, Aaditya Ramdas, Andrea Bajcsy

    Reinforcement Learning

    A Single Goal is All You Need: Skills and Exploration Emerge from Contrastive RL without Rewards, Demonstrations, or Subgoals

    Authors: Grace Liu, Michael Tang, Benjamin Eysenbach

    Diffusing States and Matching Scores: A New Framework for Imitation Learning

    Authors: Runzhe Wu, Yiding Chen, Gokul Swamy, Kianté Brantley, Wen Sun

    Efficient Imitation under Misspecification

    Authors: Nicolas Espinosa-dice, Sanjiban Choudhury, Wen Sun, Gokul Swamy

    Pairwise Elimination with Instance-Dependent Guarantees for Bandits with Cost Subsidy

    Authors: Ishank Juneja, Carlee Joe-wong, Osman Yagan

    Regressing the Relative Future: Efficient Policy Optimization for Multi-turn RLHF

    Authors: Zhaolin Gao, Wenhao Zhan, Jonathan Daniel Chang, Gokul Swamy, Kianté Brantley, Jason D. Lee, Wen Sun

    Reinforcement learning with combinatorial actions for coupled restless bandits

    Authors: Lily Xu, Bryan Wilder, Elias Boutros Khalil, Milind Tambe

    Value-Incentivized Preference Optimization: A Unified Approach to Online and Offline RLHF

    Authors: Shicong Cen, Jincheng Mei, Katayoon Goshvadi, Hanjun Dai, Tong Yang, Sherry Yang, Dale Schuurmans, Yuejie Chi, Bo Dai

    Transfer Learning, Meta Learning, And Lifelong Learning

    Many-Objective Multi-Solution Transport

    Authors: Ziyue Li, Tian Li, Virginia Smith, Jeff Bilmes, Tianyi Zhou

    pMoE: Prompting Diverse Experts Together Wins More in Visual Adaptation

    Authors: Shentong Mo, Xufang Luo, Dongsheng Li

    Unsupervised, Self-supervised, Semi-supervised, And Supervised Representation Learning

    Learning Representations of Intermittent Temporal Latent Process

    Authors: Yuke Li, Yujia Zheng, Guangyi Chen, Kun Zhang, Heng Huang

    Memory Mosaics

    Authors: Jianyu Zhang, Niklas Nolte, Ranajoy Sadhukhan, Beidi Chen, Leon Bottou

    MetaOOD: Automatic Selection of OOD Detection Models

    Authors: Yuehan Qin, Yichi Zhang, Yi Nian, Xueying Ding, Yue Zhao

    Repetition Improves Language Model Embeddings

    Authors: Jacob Mitchell Springer, Suhas Kotha, Daniel Fried, Graham Neubig, Aditi Raghunathan

    Synergy Between Sufficient Changes and Sparse Mixing Procedure for Disentangled Representation Learning

    Authors: Zijian Li, Shunxing Fan, Yujia Zheng, Ignavier Ng, Shaoan Xie, Guangyi Chen, Xinshuai Dong, Ruichu Cai, Kun Zhang

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleNativePHP for Mobile v1 — Launching May 2
    Next Article Reimagining Legacy Systems with AI: Why We’re Building the Future

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    June 8, 2025
    Machine Learning

    Meet BioReason: The World’s First Reasoning Model in Biology that Enables AI to Reason about Genomics like a Biology Expert

    June 8, 2025
    Leave A Reply Cancel Reply

    For security, use of Google's reCAPTCHA service is required which is subject to the Google Privacy Policy and Terms of Use.

    Continue Reading

    Warhammer 40,000: Space Marine is remastered for Xbox and PC, and it’s coming day one to Game Pass

    News & Updates

    Next.js vs. Traditional React: What Businesses Need to Know

    Web Development

    CVE-2025-4429 – “Gearside Developer Dashboard WordPress Plugin Reflected Cross-Site Scripting”

    Common Vulnerabilities and Exposures (CVEs)

    Google DeepMind at NeurIPS 2023

    Artificial Intelligence

    Highlights

    Critical Vulnerabilities in Quick Agent Software Expose Ricoh MFPs to Remote Attacks

    April 29, 2025

    Critical Vulnerabilities in Quick Agent Software Expose Ricoh MFPs to Remote Attacks

    The Japan Computer Emergency Response Team Coordination Center (JPCERT/CC) has issued an alert regarding multiple critical vulnerabilities found in Quick Agent, a Windows application developed by SIOS …
    Read more

    Published Date:
    Apr 29, 2025 (3 hours, 22 minutes ago)

    Vulnerabilities has been mentioned in this article.

    CVE-2025-31144

    CVE-2025-27937

    CVE-2025-26692

    CVE-2025-26506

    CVE-2024-47939

    element

    April 7, 2025

    New Witcher 3 Xbox controllers look like they came straight out of the game

    May 23, 2025

    Microsoft confirms limited Microsoft 365 app support on Windows 10 after October 2025

    May 15, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.