Most Cited ICML 2025 "speech-mesh representation" Papers

3,340 papers found • Page 1 of 17

#1

WorldSimBench: Towards Video Generation Models as World Simulators

Yiran Qin, Zhelun Shi, Jiwen Yu et al.

ICML 2025arXiv:2410.18072
842
citations
#2

SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training

Tianzhe Chu, Yuexiang Zhai, Jihan Yang et al.

ICML 2025arXiv:2501.17161
442
citations
#3

From Crowdsourced Data to High-quality Benchmarks: Arena-Hard and Benchbuilder Pipeline

Tianle Li, Wei-Lin Chiang, Evan Frick et al.

ICML 2025arXiv:2406.11939
357
citations
#4

rStar-Math: Small LLMs Can Master Math Reasoning with Self-Evolved Deep Thinking

Xinyu Guan, Li Lyna Zhang, Yifei Liu et al.

ICML 2025oralarXiv:2501.04519
268
citations
#5

SparseVLM: Visual Token Sparsification for Efficient Vision-Language Model Inference

Yuan Zhang, Chun-Kai Fan, Junpeng Ma et al.

ICML 2025arXiv:2410.04417
214
citations
#6

Learning to (Learn at Test Time): RNNs with Expressive Hidden States

Yu Sun, Xinhao Li, Karan Dalal et al.

ICML 2025spotlightarXiv:2407.04620
199
citations
#7

LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding

Xiaoqian Shen, Yunyang Xiong, Changsheng Zhao et al.

ICML 2025oralarXiv:2410.17434
184
citations
#8

Aguvis: Unified Pure Vision Agents for Autonomous GUI Interaction

Yiheng Xu, Zekun Wang, Junli Wang et al.

ICML 2025arXiv:2412.04454
182
citations
#9

MimicMotion: High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance

Yuang Zhang, Jiaxi Gu, Li-Wen Wang et al.

ICML 2025oralarXiv:2406.19680
161
citations
#10

Training Software Engineering Agents and Verifiers with SWE-Gym

Jiayi Pan, Xingyao Wang, Graham Neubig et al.

ICML 2025arXiv:2412.21139
156
citations
#11

Layer by Layer: Uncovering Hidden Representations in Language Models

Oscar Skean, Md Rifat Arefin, Dan Zhao et al.

ICML 2025oralarXiv:2502.02013
145
citations
#12

AdvPrompter: Fast Adaptive Adversarial Prompting for LLMs

Anselm Paulus, Arman Zharmagambetov, Chuan Guo et al.

ICML 2025arXiv:2404.16873
132
citations
#13

Imagine While Reasoning in Space: Multimodal Visualization-of-Thought

Chengzu Li, Wenshan Wu, Huanyu Zhang et al.

ICML 2025arXiv:2501.07542
131
citations
#14

DINO-WM: World Models on Pre-trained Visual Features enable Zero-shot Planning

Gaoyue Zhou, Hengkai Pan, Yann LeCun et al.

ICML 2025oralarXiv:2411.04983
126
citations
#15

How Far Is Video Generation from World Model: A Physical Law Perspective

Bingyi Kang, Yang Yue, Rui Lu et al.

ICML 2025arXiv:2411.02385
126
citations
#16

Hi Robot: Open-Ended Instruction Following with Hierarchical Vision-Language-Action Models

Lucy Xiaoyang Shi, brian ichter, Michael Equi et al.

ICML 2025arXiv:2502.19417
120
citations
#17

A General Framework for Inference-time Scaling and Steering of Diffusion Models

Raghav Singhal, Zachary Horvitz, Ryan Teehan et al.

ICML 2025arXiv:2501.06848
119
citations
#18

Taming Rectified Flow for Inversion and Editing

Jiangshan Wang, Junfu Pu, Zhongang Qi et al.

ICML 2025arXiv:2411.04746
119
citations
#19

AxBench: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders

Zhengxuan Wu, Aryaman Arora, Atticus Geiger et al.

ICML 2025spotlightarXiv:2501.17148
118
citations
#20

Free Process Rewards without Process Labels

Lifan Yuan, Wendi Li, Huayu Chen et al.

ICML 2025arXiv:2412.01981
117
citations
#21

ZebraLogic: On the Scaling Limits of LLMs for Logical Reasoning

Yuchen Lin, Ronan Le Bras, Kyle Richardson et al.

ICML 2025arXiv:2502.01100
116
citations
#22

Agent-as-a-Judge: Evaluate Agents with Agents

Mingchen Zhuge, Changsheng Zhao, Dylan Ashley et al.

ICML 2025arXiv:2410.10934
114
citations
#23

Video Prediction Policy: A Generalist Robot Policy with Predictive Visual Representations

Yucheng Hu, Yanjiang Guo, Pengchao Wang et al.

ICML 2025spotlightarXiv:2412.14803
113
citations
#24

Freeze-Omni: A Smart and Low Latency Speech-to-speech Dialogue Model with Frozen LLM

Xiong Wang, Yangze Li, Chaoyou Fu et al.

ICML 2025arXiv:2411.00774
112
citations
#25

SANA 1.5: Efficient Scaling of Training-Time and Inference-Time Compute in Linear Diffusion Transformer

Enze Xie, Junsong Chen, Yuyang Zhao et al.

ICML 2025arXiv:2501.18427
111
citations
#26

Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs

Jan Betley, Daniel Tan, Niels Warncke et al.

ICML 2025oralarXiv:2502.17424
108
citations
#27

DPO Meets PPO: Reinforced Token Optimization for RLHF

Han Zhong, Zikang Shan, Guhao Feng et al.

ICML 2025spotlightarXiv:2404.18922
106
citations
#28

MedXpertQA: Benchmarking Expert-Level Medical Reasoning and Understanding

Yuxin Zuo, Shang Qu, Yifei Li et al.

ICML 2025arXiv:2501.18362
105
citations
#29

OR-Bench: An Over-Refusal Benchmark for Large Language Models

Jiaxing Cui, Wei-Lin Chiang, Ion Stoica et al.

ICML 2025arXiv:2405.20947
104
citations
#30

Windows Agent Arena: Evaluating Multi-Modal OS Agents at Scale

Rogerio Bonatti, Dan Zhao, Francesco Bonacci et al.

ICML 2025arXiv:2409.08264
102
citations
#31

Train for the Worst, Plan for the Best: Understanding Token Ordering in Masked Diffusions

Jaeyeon Kim, Kulin Shah, Vasilis Kontonis et al.

ICML 2025oralarXiv:2502.06768
102
citations
#32

Can MLLMs Reason in Multimodality? EMMA: An Enhanced MultiModal ReAsoning Benchmark

Yunzhuo Hao, Jiawei Gu, Huichen Wang et al.

ICML 2025oralarXiv:2501.05444
100
citations
#33

PaperBench: Evaluating AI’s Ability to Replicate AI Research

Giulio Starace, Oliver Jaffe, Dane Sherburn et al.

ICML 2025oralarXiv:2504.01848
99
citations
#34

EmbodiedBench: Comprehensive Benchmarking Multi-modal Large Language Models for Vision-Driven Embodied Agents

Rui Yang, Hanyang(Jeremy) Chen, Junyu Zhang et al.

ICML 2025oralarXiv:2502.09560
98
citations
#35

Learning Smooth and Expressive Interatomic Potentials for Physical Property Prediction

Xiang Fu, Brandon Wood, Luis Barroso-Luque et al.

ICML 2025oralarXiv:2502.12147
96
citations
#36

Theoretical guarantees on the best-of-n alignment policy

Ahmad Beirami, Alekh Agarwal, Jonathan Berant et al.

ICML 2025arXiv:2401.01879
95
citations
#37

Mind Your Step (by Step): Chain-of-Thought can Reduce Performance on Tasks where Thinking Makes Humans Worse

Ryan Liu, Jiayi Geng, Addison J. Wu et al.

ICML 2025arXiv:2410.21333
95
citations
#38

Weak-to-Strong Jailbreaking on Large Language Models

Xuandong Zhao, Xianjun Yang, Tianyu Pang et al.

ICML 2025arXiv:2401.17256
95
citations
#39

MME-CoT: Benchmarking Chain-of-Thought in Large Multimodal Models for Reasoning Quality, Robustness, and Efficiency

Dongzhi Jiang, Renrui Zhang, Ziyu Guo et al.

ICML 2025arXiv:2502.09621
94
citations
#40

RLEF: Grounding Code LLMs in Execution Feedback with Reinforcement Learning

Jonas Gehring, Kunhao Zheng, Jade Copet et al.

ICML 2025spotlightarXiv:2410.02089
91
citations
#41

Agent Workflow Memory

Zhiruo Wang, Jiayuan Mao, Daniel Fried et al.

ICML 2025arXiv:2409.07429
90
citations
#42

Forest-of-Thought: Scaling Test-Time Compute for Enhancing LLM Reasoning

Zhenni Bi, Kai Han, Chuanjian Liu et al.

ICML 2025arXiv:2412.09078
89
citations
#43

Audio Flamingo 2: An Audio-Language Model with Long-Audio Understanding and Expert Reasoning Abilities

Sreyan Ghosh, Zhifeng Kong, Sonal Kumar et al.

ICML 2025arXiv:2503.03983
88
citations
#44

TabICL: A Tabular Foundation Model for In-Context Learning on Large Data

Jingang QU, David Holzmüller, Gael Varoquaux et al.

ICML 2025arXiv:2502.05564
84
citations
#45

Plan-and-Act: Improving Planning of Agents for Long-Horizon Tasks

Lutfi Erdogan, Hiroki Furuta, Sehoon Kim et al.

ICML 2025arXiv:2503.09572
84
citations
#46

Multi-agent Architecture Search via Agentic Supernet

Guibin Zhang, Luyang Niu, Junfeng Fang et al.

ICML 2025oralarXiv:2502.04180
83
citations
#47

Reward-Guided Speculative Decoding for Efficient LLM Reasoning

Baohao Liao, Yuhui Xu, Hanze Dong et al.

ICML 2025arXiv:2501.19324
77
citations
#48

Sundial: A Family of Highly Capable Time Series Foundation Models

Yong Liu, Guo Qin, Zhiyuan Shi et al.

ICML 2025oralarXiv:2502.00816
76
citations
#49

Towards World Simulator: Crafting Physical Commonsense-Based Benchmark for Video Generation

Fanqing Meng, Jiaqi Liao, Xinyu Tan et al.

ICML 2025arXiv:2410.05363
76
citations
#50

Cradle: Empowering Foundation Agents towards General Computer Control

Weihao Tan, Wentao Zhang, Xinrun Xu et al.

ICML 2025arXiv:2403.03186
75
citations
#51

Diffusion Adversarial Post-Training for One-Step Video Generation

Shanchuan Lin, Xin Xia, Yuxi Ren et al.

ICML 2025arXiv:2501.08316
75
citations
#52

History-Guided Video Diffusion

Kiwhan Song, Boyuan Chen, Max Simchowitz et al.

ICML 2025oralarXiv:2502.06764
75
citations
#53

Scaling Test-Time Compute Without Verification or RL is Suboptimal

Amrith Setlur, Nived Rajaraman, Sergey Levine et al.

ICML 2025spotlightarXiv:2502.12118
73
citations
#54

Training Deep Learning Models with Norm-Constrained LMOs

Thomas Pethick, Wanyun Xie, Kimon Antonakopoulos et al.

ICML 2025spotlightarXiv:2502.07529
72
citations
#55

XAttention: Block Sparse Attention with Antidiagonal Scoring

Ruyi Xu, Guangxuan Xiao, Haofeng Huang et al.

ICML 2025arXiv:2503.16428
71
citations
#56

Moirai-MoE: Empowering Time Series Foundation Models with Sparse Mixture of Experts

Xu Liu, Juncheng Liu, Gerald Woo et al.

ICML 2025arXiv:2410.10469
71
citations
#57

HealthGPT: A Medical Large Vision-Language Model for Unifying Comprehension and Generation via Heterogeneous Knowledge Adaptation

Tianwei Lin, Wenqiao Zhang, Sijing Li et al.

ICML 2025spotlightarXiv:2502.09838
69
citations
#58

GuardAgent: Safeguard LLM Agents via Knowledge-Enabled Reasoning

Zhen Xiang, Linzhi Zheng, Yanjie Li et al.

ICML 2025
69
citations
#59

What If We Recaption Billions of Web Images with LLaMA-3?

Xianhang Li, Haoqin Tu, Mude Hui et al.

ICML 2025arXiv:2406.08478
69
citations
#60

T1: Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling

Zhenyu Hou, Xin Lv, Rui Lu et al.

ICML 2025arXiv:2501.11651
68
citations
#61

VideoJAM: Joint Appearance-Motion Representations for Enhanced Motion Generation in Video Models

Hila Chefer, Uriel Singer, Amit Zohar et al.

ICML 2025oralarXiv:2502.02492
68
citations
#62

LMRL Gym: Benchmarks for Multi-Turn Reinforcement Learning with Language Models

Marwa Abdulhai, Isadora White, Charlie Snell et al.

ICML 2025oralarXiv:2311.18232
67
citations
#63

SageAttention2: Efficient Attention with Thorough Outlier Smoothing and Per-thread INT4 Quantization

Jintao Zhang, Haofeng Huang, Pengle Zhang et al.

ICML 2025arXiv:2411.10958
66
citations
#64

SWE-Lancer: Can Frontier LLMs Earn $1 Million from Real-World Freelance Software Engineering?

Samuel Miserendino, Michele Wang, Tejal Patwardhan et al.

ICML 2025oralarXiv:2502.12115
66
citations
#65

ShadowKV: KV Cache in Shadows for High-Throughput Long-Context LLM Inference

Hanshi Sun, Li-Wen Chang, Wenlei Bao et al.

ICML 2025spotlightarXiv:2410.21465
65
citations
#66

RE-Bench: Evaluating Frontier AI R&D Capabilities of Language Model Agents against Human Experts

Hjalmar Wijk, Tao Lin, Joel Becker et al.

ICML 2025spotlightarXiv:2411.15114
65
citations
#67

Inductive Moment Matching

Linqi (Alex) Zhou, Stefano Ermon, Jiaming Song

ICML 2025oralarXiv:2503.07565
65
citations
#68

Flow Q-Learning

Seohong Park, Qiyang Li, Sergey Levine

ICML 2025arXiv:2502.02538
65
citations
#69

Fast Video Generation with Sliding Tile Attention

Peiyuan Zhang, Yongqi Chen, Runlong Su et al.

ICML 2025oralarXiv:2502.04507
64
citations
#70

An analytic theory of creativity in convolutional diffusion models

Mason Kamb, Surya Ganguli

ICML 2025oralarXiv:2412.20292
64
citations
#71

MATH-Perturb: Benchmarking LLMs' Math Reasoning Abilities against Hard Perturbations

Kaixuan Huang, Jiacheng Guo, Zihao Li et al.

ICML 2025arXiv:2502.06453
63
citations
#72

KernelBench: Can LLMs Write Efficient GPU Kernels?

Anne Ouyang, Simon Guo, Simran Arora et al.

ICML 2025arXiv:2502.10517
63
citations
#73

MM-RLHF: The Next Step Forward in Multimodal LLM Alignment

Yi-Fan Zhang, Tao Yu, Haochen Tian et al.

ICML 2025arXiv:2502.10391
63
citations
#74

Automatically Interpreting Millions of Features in Large Language Models

Gonçalo Paulo, Alex Mallen, Caden Juang et al.

ICML 2025arXiv:2410.13928
61
citations
#75

AutoML-Agent: A Multi-Agent LLM Framework for Full-Pipeline AutoML

Patara Trirat, Wonyong Jeong, Sung Ju Hwang

ICML 2025arXiv:2410.02958
60
citations
#76

Learning to Plan & Reason for Evaluation with Thinking-LLM-as-a-Judge

Swarnadeep Saha, Xian Li, Marjan Ghazvininejad et al.

ICML 2025arXiv:2501.18099
60
citations
#77

Normalizing Flows are Capable Generative Models

Shuangfei Zhai, Ruixiang Zhang, Preetum Nakkiran et al.

ICML 2025oralarXiv:2412.06329
59
citations
#78

SAEBench: A Comprehensive Benchmark for Sparse Autoencoders in Language Model Interpretability

Adam Karvonen, Can Rager, Johnny Lin et al.

ICML 2025arXiv:2503.09532
58
citations
#79

Learning Multi-Level Features with Matryoshka Sparse Autoencoders

Bart Bussmann, Noa Nabeshima, Adam Karvonen et al.

ICML 2025arXiv:2503.17547
58
citations
#80

NoLiMa: Long-Context Evaluation Beyond Literal Matching

Ali Modarressi, Hanieh Deilamsalehy, Franck Dernoncourt et al.

ICML 2025arXiv:2502.05167
57
citations
#81

Are Sparse Autoencoders Useful? A Case Study in Sparse Probing

Subhash Kantamneni, Josh Engels, Senthooran Rajamanoharan et al.

ICML 2025arXiv:2502.16681
56
citations
#82

VinePPO: Refining Credit Assignment in RL Training of LLMs

Amirhossein Kazemnejad, Milad Aghajohari, Eva Portelance et al.

ICML 2025arXiv:2410.01679
56
citations
#83

VisionTS: Visual Masked Autoencoders Are Free-Lunch Zero-Shot Time Series Forecasters

Mouxiang Chen, Lefei Shen, Zhuo Li et al.

ICML 2025arXiv:2408.17253
56
citations
#84

Long-Term TalkingFace Generation via Motion-Prior Conditional Diffusion Model

SHEN FEI, Cong Wang, Junyao Gao et al.

ICML 2025oralarXiv:2502.09533
53
citations
#85

Organize the Web: Constructing Domains Enhances Pre-Training Data Curation

Alexander Wettig, Kyle Lo, Sewon Min et al.

ICML 2025arXiv:2502.10341
53
citations
#86

Masked Autoencoders Are Effective Tokenizers for Diffusion Models

Hao Chen, Yujin Han, Fangyi Chen et al.

ICML 2025spotlightarXiv:2502.03444
52
citations
#87

Token Assorted: Mixing Latent and Text Tokens for Improved Language Model Reasoning

Andy (DiJia) Su, Hanlin Zhu, Yingchen Xu et al.

ICML 2025arXiv:2502.03275
52
citations
#88

Critical Tokens Matter: Token-Level Contrastive Estimation Enhances LLM’s Reasoning Capability

Zicheng Lin, Tian Liang, Jiahao Xu et al.

ICML 2025arXiv:2411.19943
51
citations
#89

EQ-VAE: Equivariance Regularized Latent Space for Improved Generative Image Modeling

Theodoros Kouzelis, Ioannis Kakogeorgiou, Spyros Gidaris et al.

ICML 2025arXiv:2502.09509
51
citations
#90

G-Designer: Architecting Multi-agent Communication Topologies via Graph Neural Networks

Guibin Zhang, Yanwei Yue, Xiangguo Sun et al.

ICML 2025spotlightarXiv:2410.11782
51
citations
#91

FlexTok: Resampling Images into 1D Token Sequences of Flexible Length

Roman Bachmann, Jesse Allardice, David Mizrahi et al.

ICML 2025arXiv:2502.13967
51
citations
#92

NETS: A Non-equilibrium Transport Sampler

Michael Albergo, Eric Vanden-Eijnden

ICML 2025arXiv:2410.02711
51
citations
#93

Monte Carlo Tree Search for Comprehensive Exploration in LLM-Based Automatic Heuristic Design

Zhi Zheng, Zhuoliang Xie, Zhenkun Wang et al.

ICML 2025arXiv:2501.08603
50
citations
#94

All-atom Diffusion Transformers: Unified generative modelling of molecules and materials

Chaitanya Joshi, Xiang Fu, Yi-Lun Liao et al.

ICML 2025arXiv:2503.03965
50
citations
#95

ReFocus: Visual Editing as a Chain of Thought for Structured Image Understanding

Xingyu Fu, Minqian Liu, Zhengyuan Yang et al.

ICML 2025arXiv:2501.05452
49
citations
#96

EraseAnything: Enabling Concept Erasure in Rectified Flow Transformers

Daiheng Gao, Shilin Lu, Wenbo Zhou et al.

ICML 2025arXiv:2412.20413
49
citations
#97

Antidote: Post-fine-tuning Safety Alignment for Large Language Models against Harmful Fine-tuning Attack

Tiansheng Huang, Gautam Bhattacharya, Pratik Joshi et al.

ICML 2025
48
citations
#98

STAIR: Improving Safety Alignment with Introspective Reasoning

Yichi Zhang, Siyuan Zhang, Yao Huang et al.

ICML 2025oralarXiv:2502.02384
48
citations
#99

AnyEdit: Edit Any Knowledge Encoded in Language Models

Houcheng Jiang, Junfeng Fang, Ningyu Zhang et al.

ICML 2025arXiv:2502.05628
47
citations
#100

FlowAR: Scale-wise Autoregressive Image Generation Meets Flow Matching

Sucheng Ren, Qihang Yu, Ju He et al.

ICML 2025arXiv:2412.15205
47
citations
#101

Learn Beneficial Noise as Graph Augmentation

Siqi Huang, Yanchen Xu, Hongyuan Zhang et al.

ICML 2025arXiv:2505.19024
47
citations
#102

FlipAttack: Jailbreak LLMs via Flipping

Yue Liu, Xiaoxin He, Miao Xiong et al.

ICML 2025arXiv:2410.02832
47
citations
#103

Empirical Design in Reinforcement Learning

Andrew Patterson, Samuel F Neumann, Martha White et al.

ICML 2025arXiv:2304.01315
46
citations
#104

Which Agent Causes Task Failures and When? On Automated Failure Attribution of LLM Multi-Agent Systems

Shaokun Zhang, Ming Yin, Jieyu Zhang et al.

ICML 2025spotlightarXiv:2505.00212
46
citations
#105

Why Is Spatial Reasoning Hard for VLMs? An Attention Mechanism Perspective on Focus Areas

Shiqi Chen, Tongyao Zhu, Ruochen Zhou et al.

ICML 2025arXiv:2503.01773
46
citations
#106

UP-VLA: A Unified Understanding and Prediction Model for Embodied Agent

Jianke Zhang, Yanjiang Guo, Yucheng Hu et al.

ICML 2025arXiv:2501.18867
45
citations
#107

On the Resilience of LLM-Based Multi-Agent Collaboration with Faulty Agents

Jen-Tse Huang, Jiaxu Zhou, Tailin Jin et al.

ICML 2025arXiv:2408.00989
45
citations
#108

The Surprising Effectiveness of Test-Time Training for Few-Shot Learning

Ekin Akyürek, Mehul Damani, Adam Zweiger et al.

ICML 2025arXiv:2411.07279
45
citations
#109

Context is Key: A Benchmark for Forecasting with Essential Textual Information

Andrew Williams, Arjun Ashok, Étienne Marcotte et al.

ICML 2025arXiv:2410.18959
45
citations
#110

FireFlow: Fast Inversion of Rectified Flow for Image Semantic Editing

Yingying Deng, Xiangyu He, Changwang Mei et al.

ICML 2025arXiv:2412.07517
45
citations
#111

One-Step Diffusion Policy: Fast Visuomotor Policies via Diffusion Distillation

Zhendong Wang, Max Li, Ajay Mandlekar et al.

ICML 2025arXiv:2410.21257
44
citations
#112

Thinking LLMs: General Instruction Following with Thought Generation

Tianhao Wu, Janice Lan, Weizhe Yuan et al.

ICML 2025arXiv:2410.10630
44
citations
#113

AdaWorld: Learning Adaptable World Models with Latent Actions

Shenyuan Gao, Siyuan Zhou, Yilun Du et al.

ICML 2025arXiv:2503.18938
44
citations
#114

RIFLEx: A Free Lunch for Length Extrapolation in Video Diffusion Transformers

Min Zhao, Guande He, Yixiao Chen et al.

ICML 2025oralarXiv:2502.15894
44
citations
#115

ShieldAgent: Shielding Agents via Verifiable Safety Policy Reasoning

Zhaorun Chen, Mintong Kang, Bo Li

ICML 2025arXiv:2503.22738
43
citations
#116

Orthogonal Subspace Decomposition for Generalizable AI-Generated Image Detection

Zhiyuan Yan, Jiangming Wang, Peng Jin et al.

ICML 2025oralarXiv:2411.15633
43
citations
#117

Is Best-of-N the Best of Them? Coverage, Scaling, and Optimality in Inference-Time Alignment

Audrey Huang, Adam Block, Qinghua Liu et al.

ICML 2025arXiv:2503.21878
43
citations
#118

An Architecture Search Framework for Inference-Time Techniques

Jon Saad-Falcon, Adrian Lafuente, Shlok Natarajan et al.

ICML 2025arXiv:2409.15254
43
citations
#119

CollabLLM: From Passive Responders to Active Collaborators

Shirley Wu, Michel Galley, Baolin Peng et al.

ICML 2025oralarXiv:2502.00640
43
citations
#120

STP: Self-play LLM Theorem Provers with Iterative Conjecturing and Proving

Kefan Dong, Tengyu Ma

ICML 2025arXiv:2502.00212
43
citations
#121

SpargeAttention: Accurate and Training-free Sparse Attention Accelerating Any Model Inference

Jintao Zhang, Chendong Xiang, Haofeng Huang et al.

ICML 2025arXiv:2502.18137
43
citations
#122

DiTAR: Diffusion Transformer Autoregressive Modeling for Speech Generation

Dongya Jia, Zhuo Chen, Jiawei Chen et al.

ICML 2025arXiv:2502.03930
41
citations
#123

Towards LLM Unlearning Resilient to Relearning Attacks: A Sharpness-Aware Minimization Perspective and Beyond

Chongyu Fan, jinghan jia, Yihua Zhang et al.

ICML 2025arXiv:2502.05374
41
citations
#124

Adjoint Sampling: Highly Scalable Diffusion Samplers via Adjoint Matching

Aaron Havens, Benjamin Kurt Miller, Bing Yan et al.

ICML 2025arXiv:2504.11713
41
citations
#125

MoH: Multi-Head Attention as Mixture-of-Head Attention

Peng Jin, Bo Zhu, Li Yuan et al.

ICML 2025arXiv:2410.11842
40
citations
#126

Satori: Reinforcement Learning with Chain-of-Action-Thought Enhances LLM Reasoning via Autoregressive Search

Maohao Shen, Guangtao Zeng, Zhenting Qi et al.

ICML 2025arXiv:2502.02508
40
citations
#127

The Diffusion Duality

Subham Sekhar Sahoo, Justin Deschenaux, Aaron Gokaslan et al.

ICML 2025arXiv:2506.10892
40
citations
#128

Feynman-Kac Correctors in Diffusion: Annealing, Guidance, and Product of Experts

Marta Skreta, Tara Akhound-Sadegh, Viktor Ohanesian et al.

ICML 2025spotlightarXiv:2503.02819
40
citations
#129

Position: Graph Learning Will Lose Relevance Due To Poor Benchmarks

Maya Bechler-Speicher, Ben Finkelshtein, Fabrizio Frasca et al.

ICML 2025arXiv:2502.14546
40
citations
#130

CVE-Bench: A Benchmark for AI Agents’ Ability to Exploit Real-World Web Application Vulnerabilities

Yuxuan Zhu, Antony Kellermann, Dylan Bowman et al.

ICML 2025spotlightarXiv:2503.17332
40
citations
#131

OTTER: A Vision-Language-Action Model with Text-Aware Visual Feature Extraction

Huang Huang, Fangchen Liu, Letian Fu et al.

ICML 2025arXiv:2503.03734
39
citations
#132

FlatQuant: Flatness Matters for LLM Quantization

Yuxuan Sun, Ruikang Liu, Haoli Bai et al.

ICML 2025arXiv:2410.09426
39
citations
#133

SliM-LLM: Salience-Driven Mixed-Precision Quantization for Large Language Models

Wei Huang, Haotong Qin, Yangdong Liu et al.

ICML 2025arXiv:2405.14917
39
citations
#134

Improving the Diffusability of Autoencoders

Ivan Skorokhodov, Sharath Girish, Benran Hu et al.

ICML 2025arXiv:2502.14831
39
citations
#135

SafeArena: Evaluating the Safety of Autonomous Web Agents

Ada Tur, Nicholas Meade, Xing Han Lù et al.

ICML 2025arXiv:2503.04957
39
citations
#136

Graph-constrained Reasoning: Faithful Reasoning on Knowledge Graphs with Large Language Models

Linhao Luo, Zicheng Zhao, Reza Haffari et al.

ICML 2025arXiv:2410.13080
38
citations
#137

FG-CLIP: Fine-Grained Visual and Textual Alignment

Chunyu Xie, Bin Wang, Fanjing Kong et al.

ICML 2025arXiv:2505.05071
38
citations
#138

WMAdapter: Adding WaterMark Control to Latent Diffusion Models

Hai Ci, Yiren Song, Pei Yang et al.

ICML 2025arXiv:2406.08337
38
citations
#139

SAeUron: Interpretable Concept Unlearning in Diffusion Models with Sparse Autoencoders

Bartosz Cywiński, Kamil Deja

ICML 2025arXiv:2501.18052
37
citations
#140

Which Attention Heads Matter for In-Context Learning?

Kayo Yin, Jacob Steinhardt

ICML 2025arXiv:2502.14010
37
citations
#141

Collapse or Thrive: Perils and Promises of Synthetic Data in a Self-Generating World

Joshua Kazdan, Rylan Schaeffer, Apratim Dey et al.

ICML 2025arXiv:2410.16713
37
citations
#142

Look Twice Before You Answer: Memory-Space Visual Retracing for Hallucination Mitigation in Multimodal Large Language Models

Xin Zou, Yizhou WANG, Yibo Yan et al.

ICML 2025arXiv:2410.03577
37
citations
#143

Learning to Route LLMs with Confidence Tokens

Yu-Neng Chuang, Prathusha Sarma, Parikshit Gopalan et al.

ICML 2025arXiv:2410.13284
36
citations
#144

AutoEval Done Right: Using Synthetic Data for Model Evaluation

Pierre Boyeau, Anastasios Angelopoulos, Tianle Li et al.

ICML 2025arXiv:2403.07008
36
citations
#145

Time-VLM: Exploring Multimodal Vision-Language Models for Augmented Time Series Forecasting

Siru Zhong, Weilin Ruan, Ming Jin et al.

ICML 2025oralarXiv:2502.04395
36
citations
#146

Transolver++: An Accurate Neural Solver for PDEs on Million-Scale Geometries

HUAKUN LUO, Haixu Wu, Hang Zhou et al.

ICML 2025arXiv:2502.02414
36
citations
#147

VideoRoPE: What Makes for Good Video Rotary Position Embedding?

Xilin Wei, Xiaoran Liu, Yuhang Zang et al.

ICML 2025oralarXiv:2502.05173
36
citations
#148

RepoAudit: An Autonomous LLM-Agent for Repository-Level Code Auditing

Jinyao Guo, Chengpeng Wang, Xiangzhe Xu et al.

ICML 2025arXiv:2501.18160
35
citations
#149

Detecting Strategic Deception with Linear Probes

Nicholas Goldowsky-Dill, Bilal Chughtai, Stefan Heimersheim et al.

ICML 2025arXiv:2502.03407
35
citations
#150

SAM2Act: Integrating Visual Foundation Model with A Memory Architecture for Robotic Manipulation

Haoquan Fang, Markus Grotz, Wilbert Pumacay et al.

ICML 2025arXiv:2501.18564
35
citations
#151

PepTune: De Novo Generation of Therapeutic Peptides with Multi-Objective-Guided Discrete Diffusion

Sophia Tang, Yinuo Zhang, Pranam Chatterjee, PhD

ICML 2025arXiv:2412.17780
35
citations
#152

DeFoG: Discrete Flow Matching for Graph Generation

Yiming Qin, Manuel Madeira, Dorina Thanou et al.

ICML 2025oralarXiv:2410.04263
35
citations
#153

Test-Time Preference Optimization: On-the-Fly Alignment via Iterative Textual Feedback

Yafu Li, Xuyang Hu, Xiaoye Qu et al.

ICML 2025arXiv:2501.12895
35
citations
#154

Why Has Predicting Downstream Capabilities of Frontier AI Models with Scale Remained Elusive?

Rylan Schaeffer, Hailey Schoelkopf, Brando Miranda et al.

ICML 2025arXiv:2406.04391
35
citations
#155

Modular Duality in Deep Learning

Jeremy Bernstein, Laker Newhouse

ICML 2025arXiv:2410.21265
35
citations
#156

Massive Values in Self-Attention Modules are the Key to Contextual Knowledge Understanding

Mingyu Jin, Kai Mei, Wujiang Xu et al.

ICML 2025arXiv:2502.01563
34
citations
#157

UI-Vision: A Desktop-centric GUI Benchmark for Visual Perception and Interaction

Perampalli Shravan Nayak, Xiangru Jian, Kevin Qinghong Lin et al.

ICML 2025arXiv:2503.15661
34
citations
#158

An Analysis of Quantile Temporal-Difference Learning

Mark Rowland, Remi Munos, Mohammad Gheshlaghi Azar et al.

ICML 2025oralarXiv:2301.04462
34
citations
#159

Robust Autonomy Emerges from Self-Play

Marco Cusumano-Towner, David Hafner, Alexander Hertzberg et al.

ICML 2025arXiv:2502.03349
34
citations
#160

AffectGPT: A New Dataset, Model, and Benchmark for Emotion Understanding with Multimodal Large Language Models

Zheng Lian, Haoyu Chen, Lan Chen et al.

ICML 2025oralarXiv:2501.16566
34
citations
#161

No Task Left Behind: Isotropic Model Merging with Common and Task-Specific Subspaces

Daniel Marczak, Simone Magistri, Sebastian Cygert et al.

ICML 2025arXiv:2502.04959
34
citations
#162

On the Emergence of Position Bias in Transformers

Xinyi Wu, Yifei Wang, Stefanie Jegelka et al.

ICML 2025arXiv:2502.01951
34
citations
#163

High-Dimensional Prediction for Sequential Decision Making

Georgy Noarov, Ramya Ramalingam, Aaron Roth et al.

ICML 2025oralarXiv:2310.17651
33
citations
#164

DEFAME: Dynamic Evidence-based FAct-checking with Multimodal Experts

Tobias Braun, Mark Rothermel, Marcus Rohrbach et al.

ICML 2025oralarXiv:2412.10510
33
citations
#165

Optimizing Large Language Model Training Using FP4 Quantization

Ruizhe Wang, Yeyun Gong, Xiao Liu et al.

ICML 2025arXiv:2501.17116
33
citations
#166

The dark side of the forces: assessing non-conservative force models for atomistic machine learning

Filippo Bigi, Marcel Langer, Michele Ceriotti

ICML 2025oralarXiv:2412.11569
33
citations
#167

Diverging Preferences: When do Annotators Disagree and do Models Know?

Michael Zhang, Zhilin Wang, Jena Hwang et al.

ICML 2025arXiv:2410.14632
33
citations
#168

Proposer-Agent-Evaluator (PAE): Autonomous Skill Discovery For Foundation Model Internet Agents

Yifei Zhou, Qianlan Yang, Kaixiang Lin et al.

ICML 2025arXiv:2412.13194
32
citations
#169

Archetypal SAE: Adaptive and Stable Dictionary Learning for Concept Extraction in Large Vision Models

Thomas Fel, Ekdeep Singh Lubana, Jacob Prince et al.

ICML 2025arXiv:2502.12892
32
citations
#170

Automated Red Teaming with GOAT: the Generative Offensive Agent Tester

Maya Pavlova, Erik Brinkman, Krithika Iyer et al.

ICML 2025arXiv:2410.01606
32
citations
#171

The Geometry of Refusal in Large Language Models: Concept Cones and Representational Independence

Tom Wollschläger, Jannes Elstner, Simon Geisler et al.

ICML 2025arXiv:2502.17420
32
citations
#172

Steer LLM Latents for Hallucination Detection

Seongheon Park, Xuefeng Du, Min-Hsuan Yeh et al.

ICML 2025arXiv:2503.01917
31
citations
#173

SepLLM: Accelerate Large Language Models by Compressing One Segment into One Separator

Guoxuan Chen, Han Shi, jiawei li et al.

ICML 2025arXiv:2412.12094
31
citations
#174

Understanding Chain-of-Thought in LLMs through Information Theory

Jean-Francois Ton, Muhammad Faaiz Taufiq, Yang Liu

ICML 2025arXiv:2411.11984
31
citations
#175

Sample, Scrutinize and Scale: Effective Inference-Time Search by Scaling Verification

Eric Zhao, Pranjal Awasthi, Sreenivas Gollapudi

ICML 2025arXiv:2502.01839
31
citations
#176

UGPhysics: A Comprehensive Benchmark for Undergraduate Physics Reasoning with Large Language Models

Xin Xu, Qiyun Xu, Tong Xiao et al.

ICML 2025arXiv:2502.00334
31
citations
#177

Efficient Online Reinforcement Learning for Diffusion Policy

Haitong Ma, Tianyi Chen, Kai Wang et al.

ICML 2025arXiv:2502.00361
31
citations
#178

FourierMamba: Fourier Learning Integration with State Space Models for Image Deraining

Dong Li, Yidi Liu, Xueyang Fu et al.

ICML 2025oralarXiv:2405.19450
31
citations
#179

Overtrained Language Models Are Harder to Fine-Tune

Jacob Mitchell Springer, Sachin Goyal, Kaiyue Wen et al.

ICML 2025arXiv:2503.19206
31
citations
#180

GeoPixel: Pixel Grounding Large Multimodal Model in Remote Sensing

Akashah Shabbir, Ilmuz Zaman Mohammed Zumri, Mohammed Bennamoun et al.

ICML 2025arXiv:2501.13925
31
citations
#181

KABB: Knowledge-Aware Bayesian Bandits for Dynamic Expert Coordination in Multi-Agent Systems

Jusheng Zhang, Zimeng Huang, Yijia Fan et al.

ICML 2025arXiv:2502.07350
30
citations
#182

Locate-then-edit for Multi-hop Factual Recall under Knowledge Editing

Zhuoran Zhang, Yongxiang Li, Zijian Kan et al.

ICML 2025arXiv:2410.06331
30
citations
#183

Distillation Scaling Laws

Dan Busbridge, Amitis Shidani, Floris Weers et al.

ICML 2025arXiv:2502.08606
30
citations
#184

How to set AdamW's weight decay as you scale model and dataset size

Xi Wang, Laurence Aitchison

ICML 2025arXiv:2405.13698
30
citations
#185

Over-Tokenized Transformer: Vocabulary is Generally Worth Scaling

Hongzhi Huang, Defa Zhu, Banggu Wu et al.

ICML 2025arXiv:2501.16975
30
citations
#186

Diving into Self-Evolving Training for Multimodal Reasoning

Wei Liu, Junlong Li, Xiwen Zhang et al.

ICML 2025arXiv:2412.17451
29
citations
#187

AdvAgent: Controllable Blackbox Red-teaming on Web Agents

Chejian Xu, Mintong Kang, Jiawei Zhang et al.

ICML 2025arXiv:2410.17401
29
citations
#188

MARS: Unleashing the Power of Variance Reduction for Training Large Models

Huizhuo Yuan, Yifeng Liu, Shuang Wu et al.

ICML 2025arXiv:2411.10438
29
citations
#189

LMAct: A Benchmark for In-Context Imitation Learning with Long Multimodal Demonstrations

Anian Ruoss, Fabio Pardo, Harris Chan et al.

ICML 2025arXiv:2412.01441
29
citations
#190

Training Dynamics of In-Context Learning in Linear Attention

Yedi Zhang, Aaditya Singh, Peter Latham et al.

ICML 2025spotlightarXiv:2501.16265
29
citations
#191

Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models

Samira Abnar, Harshay Shah, Dan Busbridge et al.

ICML 2025arXiv:2501.12370
29
citations
#192

Orient Anything: Learning Robust Object Orientation Estimation from Rendering 3D Models

Zehan Wang, Ziang Zhang, Tianyu Pang et al.

ICML 2025arXiv:2412.18605
29
citations
#193

RUN: Reversible Unfolding Network for Concealed Object Segmentation

Chunming He, Rihan Zhang, Fengyang Xiao et al.

ICML 2025arXiv:2501.18783
29
citations
#194

Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale

Fan Zhou, Zengzhi Wang, Qian Liu et al.

ICML 2025arXiv:2409.17115
29
citations
#195

From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models

Etowah Adams, Liam Bai, Minji Lee et al.

ICML 2025spotlight
28
citations
#196

Subspace Optimization for Large Language Models with Convergence Guarantees

Yutong He, Pengrui Li, Yipeng Hu et al.

ICML 2025arXiv:2410.11289
28
citations
#197

Orthus: Autoregressive Interleaved Image-Text Generation with Modality-Specific Heads

Siqi Kou, Jiachun Jin, Zhihong Liu et al.

ICML 2025arXiv:2412.00127
28
citations
#198

AI for Global Climate Cooperation: Modeling Global Climate Negotiations, Agreements, and Long-Term Cooperation in RICE-N

Tianyu Zhang, Andrew Williams, Phillip Wozny et al.

ICML 2025arXiv:2208.07004
28
citations
#199

Star Attention: Efficient LLM Inference over Long Sequences

Shantanu Acharya, Fei Jia, Boris Ginsburg

ICML 2025arXiv:2411.17116
28
citations
#200

LLM-SRBench: A New Benchmark for Scientific Equation Discovery with Large Language Models

Parshin Shojaee, Ngoc Hieu Nguyen, Kazem Meidani et al.

ICML 2025oralarXiv:2504.10415
28
citations
PreviousNext