ICML Spotlight Papers
406 papers found • Page 4 of 9
Position: Formal Mathematical Reasoning—A New Frontier in AI
Kaiyu Yang, Gabriel Poesia, Jingxuan He et al.
Position: General Intelligence Requires Reward-based Pretraining
Seungwook Han, Jyothish Pari, Samuel Gershman et al.
Position: Human Baselines in Model Evaluations Need Rigor and Transparency (With Recommendations & Reporting Checklist)
Kevin Wei, Patricia Paskov, Sunishchal Dev et al.
Position: In-House Evaluation Is Not Enough. Towards Robust Third-Party Evaluation and Flaw Disclosure for General-Purpose AI
Shayne Longpre, Kevin Klyman, Ruth Elisabeth Appel et al.
Position: Language model developers should report train-test overlap
Andy Zhang, Kevin Klyman, Yifan Mai et al.
Position: Rethinking LLM Bias Probing Using Lessons from the Social Sciences
Kirsten Morehouse, Siddharth Swaroop, Weiwei Pan
Position: The Categorization of Race in ML is a Flawed Premise
Miriam Doh, Benedikt Höltgen, Piera Riccio et al.
Position: We Can’t Understand AI Using our Existing Vocabulary
John Hewitt, Robert Geirhos, Been Kim
Position: We Need An Algorithmic Understanding of Generative AI
Oliver Eberle, Thomas McGee, Hamza Giaffar et al.
Prediction models that learn to avoid missing values
Lena Stempfle, Anton Matsson, Newton Mwai et al.
Primal-Dual Neural Algorithmic Reasoning
Yu He, Ellen Vitercik
Privacy Amplification by Structured Subsampling for Deep Differentially Private Time Series Forecasting
Jan Schuchardt, Mina Dalirrooyfard, Jed Guzelkabaagac et al.
Probabilistic Factorial Experimental Design for Combinatorial Interventions
Divya Shyamal, Jiaqi Zhang, Caroline Uhler
Procurement Auctions via Approximately Optimal Submodular Optimization
Yuan Deng, Amin Karbasi, Vahab Mirrokni et al.
Provable Benefits of Unsupervised Pre-training and Transfer Learning via Single-Index Models
Taj Jones-McCormick, Aukosh Jagannath, Subhabrata Sen
RAPID: Long-Context Inference with Retrieval-Augmented Speculative Decoding
Guanzheng Chen, Qilong Feng, Jinjie Ni et al.
Rapid Overfitting of Multi-Pass SGD in Stochastic Convex Optimization
Shira Vansover-Hager, Tomer Koren, Roi Livni
Raptor: Scalable Train-Free Embeddings for 3D Medical Volumes Leveraging Pretrained 2D Foundation Models
Ulzee An, Moonseong Jeong, Simon Lee et al.
RE-Bench: Evaluating Frontier AI R&D Capabilities of Language Model Agents against Human Experts
Hjalmar Wijk, Tao Lin, Joel Becker et al.
Reducing Variance of Stochastic Optimization for Approximating Nash Equilibria in Normal-Form Games
Linjian Meng, Wubing Chen, Wenbin Li et al.
Relational Invariant Learning for Robust Solvation Free Energy Prediction
Yeyun Chen
Re-ranking Reasoning Context with Tree Search Makes Large Vision-Language Models Stronger
Qi Yang, Chenghao Zhang, Lubin Fan et al.
ResQ: Mixed-Precision Quantization of Large Language Models with Low-Rank Residuals
Utkarsh Saxena, Sayeh Sharify, Kaushik Roy et al.
Rethink GraphODE Generalization within Coupled Dynamical System
Guancheng Wan, Zijie Huang, Wanjia Zhao et al.
Return of the Latent Space COWBOYS: Re-thinking the use of VAEs for Bayesian Optimisation of Structured Spaces
Henry Moss, Sebastian Ober, Tom Diethe
Revisiting Continuity of Image Tokens for Cross-domain Few-shot Learning
Shuai Yi, Yixiong Zou, Yuhua Li et al.
RLEF: Grounding Code LLMs in Execution Feedback with Reinforcement Learning
Jonas Gehring, Kunhao Zheng, Jade Copet et al.
Robust Automatic Modulation Classification with Fuzzy Regularization
Xinyan Liang, Ruijie Sang, Yuhua Qian et al.
Robust ML Auditing using Prior Knowledge
Jade Garcia Bourrée, Augustin Godinot, Sayan Biswas et al.
Robust Noise Attenuation via Adaptive Pooling of Transformer Outputs
Greyson Brothers
SAFE: Finding Sparse and Flat Minima to Improve Pruning
Dongyeop Lee, Kwanhee Lee, Jinseok Chung et al.
Scalable Generation of Spatial Transcriptomics from Histology Images via Whole-Slide Flow Matching
Tinglin Huang, Tianyu Liu, Mehrtash Babadi et al.
Scaling Test-Time Compute Without Verification or RL is Suboptimal
Amrith Setlur, Nived Rajaraman, Sergey Levine et al.
Scaling Trends in Language Model Robustness
Nikolaus Howe, Ian McKenzie, Oskar Hollinsworth et al.
Score-of-Mixture Training: One-Step Generative Model Training Made Simple via Score Estimation of Mixture Distributions
Tejas Jayashankar, Jongha (Jon) Ryu, Gregory Wornell
scSSL-Bench: Benchmarking Self-Supervised Learning for Single-Cell Data
Olga Ovcharenko, Florian Barkmann, Philip Toma et al.
SDP-CROWN: Efficient Bound Propagation for Neural Network Verification with Tightness of Semidefinite Programming
Hong-Ming Chiu, Hao Chen, Huan Zhang et al.
Self-supervised Masked Graph Autoencoder via Structure-aware Curriculum
Haoyang Li, Xin Wang, Zeyang Zhang et al.
ShadowKV: KV Cache in Shadows for High-Throughput Long-Context LLM Inference
Hanshi Sun, Li-Wen Chang, Wenlei Bao et al.
Sharp Generalization for Nonparametric Regression by Over-Parameterized Neural Networks: A Distribution-Free Analysis in Spherical Covariate
Yingzhen Yang
Signed Laplacians for Constrained Graph Clustering
John Stewart Fabila-Carrasco, He Sun
Soft Reasoning: Navigating Solution Spaces in Large Language Models through Controlled Embedding Exploration
Qinglin Zhu, Runcong Zhao, Hanqi Yan et al.
Sparse-pivot: Dynamic correlation clustering for node insertions
Mina Dalirrooyfard, Konstantin Makarychev, Slobodan Mitrovic
STAR: Learning Diverse Robot Skill Abstractions through Rotation-Augmented Vector Quantization
Hao Li, Qi Lv, Rui Shao et al.
Stochastic Smoothed Primal-Dual Algorithms for Nonconvex Optimization with Linear Inequality Constraints
Ruichuan Huang, Jiawei Zhang, Ahmet Alacaoglu
Stronger Neyman Regret Guarantees for Adaptive Experimental Design
Georgy Noarov, Riccardo Fogliato, Martin A Bertran et al.
TabFlex: Scaling Tabular Learning to Millions with Linear Attention
Yuchen Zeng, Tuan Dinh, Wonjun Kang et al.
Taming Knowledge Conflicts in Language Models
Gaotang Li, Yuzhong Chen, Hanghang Tong
The Berkeley Function Calling Leaderboard (BFCL): From Tool Use to Agentic Evaluation of Large Language Models
Shishir G. Patil, Huanzhi Mao, Fanjia Yan et al.
The Jailbreak Tax: How Useful are Your Jailbreak Outputs?
Kristina Nikolić, Luze Sun, Jie Zhang et al.