Most Cited ICML "spatiotemporal dependencies" Papers

5,975 papers found • Page 2 of 30

#201

SparQ Attention: Bandwidth-Efficient LLM Inference

Luka Ribar, Ivan Chelombiev, Luke Hudlass-Galley et al.

ICML 2024arXiv:2312.04985
90
citations
#202

Forest-of-Thought: Scaling Test-Time Compute for Enhancing LLM Reasoning

Zhenni Bi, Kai Han, Chuanjian Liu et al.

ICML 2025arXiv:2412.09078
89
citations
#203

MaxMin-RLHF: Alignment with Diverse Human Preferences

Souradip Chakraborty, Jiahao Qiu, Hui Yuan et al.

ICML 2024arXiv:2402.08925
88
citations
#204

MagicLens: Self-Supervised Image Retrieval with Open-Ended Instructions

Kai Zhang, Yi Luan, Hexiang Hu et al.

ICML 2024arXiv:2403.19651
88
citations
#205

Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning

Hao Zhao, Maksym Andriushchenko, Francesco Croce et al.

ICML 2024arXiv:2402.04833
88
citations
#206

Audio Flamingo 2: An Audio-Language Model with Long-Audio Understanding and Expert Reasoning Abilities

Sreyan Ghosh, Zhifeng Kong, Sonal Kumar et al.

ICML 2025arXiv:2503.03983
88
citations
#207

Human Alignment of Large Language Models through Online Preference Optimisation

Daniele Calandriello, Zhaohan Guo, REMI MUNOS et al.

ICML 2024arXiv:2403.08635
88
citations
#208

Robust CLIP: Unsupervised Adversarial Fine-Tuning of Vision Embeddings for Robust Large Vision-Language Models

Christian Schlarmann, Naman Singh, Francesco Croce et al.

ICML 2024arXiv:2402.12336
88
citations
#209

Representation Surgery for Multi-Task Model Merging

Enneng Yang, Li Shen, Zhenyi Wang et al.

ICML 2024arXiv:2402.02705
87
citations
#210

IM-3D: Iterative Multiview Diffusion and Reconstruction for High-Quality 3D Generation

Luke Melas-Kyriazi, Iro Laina, Christian Rupprecht et al.

ICML 2024arXiv:2402.08682
87
citations
#211

FlowMM: Generating Materials with Riemannian Flow Matching

Benjamin Kurt Miller, Ricky T. Q. Chen, Anuroop Sriram et al.

ICML 2024arXiv:2406.04713
87
citations
#212

Language Models with Conformal Factuality Guarantees

Christopher Mohri, Tatsunori Hashimoto

ICML 2024arXiv:2402.10978
85
citations
#213

A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts

Kuang-Huei Lee, Xinyun Chen, Hiroki Furuta et al.

ICML 2024arXiv:2402.09727
85
citations
#214

AnyTool: Self-Reflective, Hierarchical Agents for Large-Scale API Calls

YU DU, Fangyun Wei, Hongyang Zhang

ICML 2024arXiv:2402.04253
85
citations
#215

Merging Multi-Task Models via Weight-Ensembling Mixture of Experts

Anke Tang, Li Shen, Yong Luo et al.

ICML 2024arXiv:2402.00433
84
citations
#216

Scalable High-Resolution Pixel-Space Image Synthesis with Hourglass Diffusion Transformers

Katherine Crowson, Stefan Baumann, Alex Birch et al.

ICML 2024arXiv:2401.11605
84
citations
#217

Plan-and-Act: Improving Planning of Agents for Long-Horizon Tasks

Lutfi Erdogan, Hiroki Furuta, Sehoon Kim et al.

ICML 2025arXiv:2503.09572
84
citations
#218

Robust Classification via a Single Diffusion Model

Huanran Chen, Yinpeng Dong, Zhengyi Wang et al.

ICML 2024arXiv:2305.15241
84
citations
#219

TabICL: A Tabular Foundation Model for In-Context Learning on Large Data

Jingang QU, David Holzmüller, Gael Varoquaux et al.

ICML 2025arXiv:2502.05564
84
citations
#220

Evaluating Quantized Large Language Models

Shiyao Li, Xuefei Ning, Luning Wang et al.

ICML 2024arXiv:2402.18158
83
citations
#221

SceneCraft: An LLM Agent for Synthesizing 3D Scenes as Blender Code

ziniu hu, Ahmet Iscen, Aashi Jain et al.

ICML 2024arXiv:2403.01248
83
citations
#222

Multi-agent Architecture Search via Agentic Supernet

Guibin Zhang, Luyang Niu, Junfeng Fang et al.

ICML 2025oralarXiv:2502.04180
83
citations
#223

In-Context Language Learning: Architectures and Algorithms

Ekin Akyürek, Bailin Wang, Yoon Kim et al.

ICML 2024arXiv:2401.12973
83
citations
#224

A Closer Look at the Limitations of Instruction Tuning

Sreyan Ghosh, Chandra Kiran Evuru, Sonal Kumar et al.

ICML 2024arXiv:2402.05119
83
citations
#225

Guiding LLMs The Right Way: Fast, Non-Invasive Constrained Generation

Luca Beurer-Kellner, Marc Fischer, Martin Vechev

ICML 2024arXiv:2403.06988
82
citations
#226

Boximator: Generating Rich and Controllable Motions for Video Synthesis

Jiawei Wang, Yuchen Zhang, Jiaxin Zou et al.

ICML 2024arXiv:2402.01566
82
citations
#227

Video-LaVIT: Unified Video-Language Pre-training with Decoupled Visual-Motional Tokenization

Yang Jin, Zhicheng Sun, Kun Xu et al.

ICML 2024oralarXiv:2402.03161
82
citations
#228

DS-Agent: Automated Data Science by Empowering Large Language Models with Case-Based Reasoning

Siyuan Guo, Cheng Deng, Ying Wen et al.

ICML 2024arXiv:2402.17453
81
citations
#229

Low-Cost High-Power Membership Inference Attacks

Sajjad Zarifzadeh, Philippe Liu, Reza Shokri

ICML 2024arXiv:2312.03262
80
citations
#230

Position: Graph Foundation Models Are Already Here

Haitao Mao, Zhikai Chen, Wenzhuo Tang et al.

ICML 2024spotlightarXiv:2402.02216
79
citations
#231

Get More with LESS: Synthesizing Recurrence with KV Cache Compression for Efficient LLM Inference

Harry Dong, Xinyu Yang, Zhenyu Zhang et al.

ICML 2024arXiv:2402.09398
79
citations
#232

Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations

Yanda Chen, Ruiqi Zhong, Narutatsu Ri et al.

ICML 2024spotlightarXiv:2307.08678
79
citations
#233

Wukong: Towards a Scaling Law for Large-Scale Recommendation

Buyun Zhang, Liang Luo, Yuxin Chen et al.

ICML 2024arXiv:2403.02545
78
citations
#234

NExT-Chat: An LMM for Chat, Detection and Segmentation

Ao Zhang, Yuan Yao, Wei Ji et al.

ICML 2024arXiv:2311.04498
78
citations
#235

A Dynamical Model of Neural Scaling Laws

Blake Bordelon, Alexander Atanasov, Cengiz Pehlevan

ICML 2024arXiv:2402.01092
77
citations
#236

Reward-Guided Speculative Decoding for Efficient LLM Reasoning

Baohao Liao, Yuhui Xu, Hanze Dong et al.

ICML 2025arXiv:2501.19324
77
citations
#237

FiT: Flexible Vision Transformer for Diffusion Model

Zeyu Lu, ZiDong Wang, Di Huang et al.

ICML 2024spotlightarXiv:2402.12376
77
citations
#238

Sundial: A Family of Highly Capable Time Series Foundation Models

Yong Liu, Guo Qin, Zhiyuan Shi et al.

ICML 2025oralarXiv:2502.00816
76
citations
#239

video-SALMONN: Speech-Enhanced Audio-Visual Large Language Models

Guangzhi Sun, Wenyi Yu, Changli Tang et al.

ICML 2024oralarXiv:2406.15704
76
citations
#240

Towards World Simulator: Crafting Physical Commonsense-Based Benchmark for Video Generation

Fanqing Meng, Jiaqi Liao, Xinyu Tan et al.

ICML 2025arXiv:2410.05363
76
citations
#241

Rolling Diffusion Models

David Ruhe, Jonathan Heek, Tim Salimans et al.

ICML 2024oralarXiv:2402.09470
76
citations
#242

How Well Can LLMs Negotiate? NegotiationArena Platform and Analysis

Federico Bianchi, Patrick John Chia, Mert Yuksekgonul et al.

ICML 2024oralarXiv:2402.05863
76
citations
#243

Cradle: Empowering Foundation Agents towards General Computer Control

Weihao Tan, Wentao Zhang, Xinrun Xu et al.

ICML 2025arXiv:2403.03186
75
citations
#244

History-Guided Video Diffusion

Kiwhan Song, Boyuan Chen, Max Simchowitz et al.

ICML 2025oralarXiv:2502.06764
75
citations
#245

D-Flow: Differentiating through Flows for Controlled Generation

Heli Ben-Hamu, Omri Puny, Itai Gat et al.

ICML 2024arXiv:2402.14017
75
citations
#246

Watermark Stealing in Large Language Models

Nikola Jovanović, Robin Staab, Martin Vechev

ICML 2024arXiv:2402.19361
75
citations
#247

Diffusion Adversarial Post-Training for One-Step Video Generation

Shanchuan Lin, Xin Xia, Yuxi Ren et al.

ICML 2025arXiv:2501.08316
75
citations
#248

Accurate LoRA-Finetuning Quantization of LLMs via Information Retention

Haotong Qin, Xudong Ma, Xingyu Zheng et al.

ICML 2024arXiv:2402.05445
74
citations
#249

A Touch, Vision, and Language Dataset for Multimodal Alignment

Letian Fu, Gaurav Datta, Huang Huang et al.

ICML 2024arXiv:2402.13232
74
citations
#250

DITTO: Diffusion Inference-Time T-Optimization for Music Generation

Zachary Novack, Julian McAuley, Taylor Berg-Kirkpatrick et al.

ICML 2024arXiv:2401.12179
74
citations
#251

Stay on Topic with Classifier-Free Guidance

Guillaume Sanchez, Alexander Spangher, Honglu Fan et al.

ICML 2024spotlightarXiv:2306.17806
73
citations
#252

Guidance with Spherical Gaussian Constraint for Conditional Diffusion

Lingxiao Yang, Shutong Ding, Yifan Cai et al.

ICML 2024arXiv:2402.03201
73
citations
#253

SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks

Jiwon Song, Kyungseok Oh, Taesu Kim et al.

ICML 2024arXiv:2402.09025
73
citations
#254

DE-COP: Detecting Copyrighted Content in Language Models Training Data

André Duarte, Xuandong Zhao, Arlindo Oliveira et al.

ICML 2024arXiv:2402.09910
73
citations
#255

VideoPrism: A Foundational Visual Encoder for Video Understanding

Long Zhao, Nitesh Bharadwaj Gundavarapu, Liangzhe Yuan et al.

ICML 2024arXiv:2402.13217
73
citations
#256

Scaling Test-Time Compute Without Verification or RL is Suboptimal

Amrith Setlur, Nived Rajaraman, Sergey Levine et al.

ICML 2025spotlightarXiv:2502.12118
73
citations
#257

DistiLLM: Towards Streamlined Distillation for Large Language Models

Jongwoo Ko, Sungnyun Kim, Tianyi Chen et al.

ICML 2024arXiv:2402.03898
73
citations
#258

Vanilla Bayesian Optimization Performs Great in High Dimensions

Carl Hvarfner, Erik Hellsten, Luigi Nardi

ICML 2024arXiv:2402.02229
72
citations
#259

Fast Adversarial Attacks on Language Models In One GPU Minute

Vinu Sankar Sadasivan, Shoumik Saha, Gaurang Sriramanan et al.

ICML 2024arXiv:2402.15570
72
citations
#260

Training Deep Learning Models with Norm-Constrained LMOs

Thomas Pethick, Wanyun Xie, Kimon Antonakopoulos et al.

ICML 2025spotlightarXiv:2502.07529
72
citations
#261

WAVES: Benchmarking the Robustness of Image Watermarks

Bang An, Mucong Ding, Tahseen Rabbani et al.

ICML 2024arXiv:2401.08573
72
citations
#262

XAttention: Block Sparse Attention with Antidiagonal Scoring

Ruyi Xu, Guangxuan Xiao, Haofeng Huang et al.

ICML 2025arXiv:2503.16428
71
citations
#263

Learning to Model the World With Language

Jessy Lin, Yuqing Du, Olivia Watkins et al.

ICML 2024arXiv:2308.01399
71
citations
#264

Moirai-MoE: Empowering Time Series Foundation Models with Sparse Mixture of Experts

Xu Liu, Juncheng Liu, Gerald Woo et al.

ICML 2025arXiv:2410.10469
71
citations
#265

DiffDA: a Diffusion model for weather-scale Data Assimilation

Langwen Huang, Lukas Gianinazzi, Yuejiang Yu et al.

ICML 2024arXiv:2401.05932
70
citations
#266

InstructRetro: Instruction Tuning post Retrieval-Augmented Pretraining

Boxin Wang, Wei Ping, Lawrence McAfee et al.

ICML 2024arXiv:2310.07713
70
citations
#267

Accelerating Convergence of Score-Based Diffusion Models, Provably

Gen Li, Yu Huang, Timofey Efimov et al.

ICML 2024arXiv:2403.03852
70
citations
#268

Learning a Diffusion Model Policy from Rewards via Q-Score Matching

Michael Psenka, Alejandro Escontrela, Pieter Abbeel et al.

ICML 2024arXiv:2312.11752
70
citations
#269

GuardAgent: Safeguard LLM Agents via Knowledge-Enabled Reasoning

Zhen Xiang, Linzhi Zheng, Yanjie Li et al.

ICML 2025
69
citations
#270

DOGE: Domain Reweighting with Generalization Estimation

Simin Fan, Matteo Pagliardini, Martin Jaggi

ICML 2024arXiv:2310.15393
69
citations
#271

VoroNav: Voronoi-based Zero-shot Object Navigation with Large Language Model

Pengying Wu, Yao Mu, Bingxian Wu et al.

ICML 2024arXiv:2401.02695
69
citations
#272

HealthGPT: A Medical Large Vision-Language Model for Unifying Comprehension and Generation via Heterogeneous Knowledge Adaptation

Tianwei Lin, Wenqiao Zhang, Sijing Li et al.

ICML 2025spotlightarXiv:2502.09838
69
citations
#273

What If We Recaption Billions of Web Images with LLaMA-3?

Xianhang Li, Haoqin Tu, Mude Hui et al.

ICML 2025arXiv:2406.08478
69
citations
#274

Stochastic Interpolants with Data-Dependent Couplings

Michael Albergo, Mark Goldstein, Nicholas Boffi et al.

ICML 2024spotlightarXiv:2310.03725
68
citations
#275

T1: Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling

Zhenyu Hou, Xin Lv, Rui Lu et al.

ICML 2025arXiv:2501.11651
68
citations
#276

VideoJAM: Joint Appearance-Motion Representations for Enhanced Motion Generation in Video Models

Hila Chefer, Uriel Singer, Amit Zohar et al.

ICML 2025oralarXiv:2502.02492
68
citations
#277

Asymmetry in Low-Rank Adapters of Foundation Models

Jiacheng Zhu, Kristjan Greenewald, Kimia Nadjahi et al.

ICML 2024arXiv:2402.16842
68
citations
#278

Image Fusion via Vision-Language Model

Zixiang Zhao, Lilun Deng, Haowen Bai et al.

ICML 2024arXiv:2402.02235
67
citations
#279

LMRL Gym: Benchmarks for Multi-Turn Reinforcement Learning with Language Models

Marwa Abdulhai, Isadora White, Charlie Snell et al.

ICML 2025oralarXiv:2311.18232
67
citations
#280

LLM and Simulation as Bilevel Optimizers: A New Paradigm to Advance Physical Scientific Discovery

Pingchuan Ma, Johnson Tsun-Hsuan Wang, Minghao Guo et al.

ICML 2024arXiv:2405.09783
67
citations
#281

RigorLLM: Resilient Guardrails for Large Language Models against Undesired Content

Zhuowen Yuan, Zidi Xiong, Yi Zeng et al.

ICML 2024arXiv:2403.13031
67
citations
#282

SWE-Lancer: Can Frontier LLMs Earn $1 Million from Real-World Freelance Software Engineering?

Samuel Miserendino, Michele Wang, Tejal Patwardhan et al.

ICML 2025oralarXiv:2502.12115
66
citations
#283

SageAttention2: Efficient Attention with Thorough Outlier Smoothing and Per-thread INT4 Quantization

Jintao Zhang, Haofeng Huang, Pengle Zhang et al.

ICML 2025arXiv:2411.10958
66
citations
#284

Long-Tail Learning with Foundation Model: Heavy Fine-Tuning Hurts

Jiang-Xin Shi, Tong Wei, Zhi Zhou et al.

ICML 2024arXiv:2309.10019
66
citations
#285

Large Language Models Can Automatically Engineer Features for Few-Shot Tabular Learning

Sungwon Han, Jinsung Yoon, Sercan Arik et al.

ICML 2024arXiv:2404.09491
66
citations
#286

Repoformer: Selective Retrieval for Repository-Level Code Completion

Di Wu, Wasi Ahmad, Dejiao Zhang et al.

ICML 2024arXiv:2403.10059
65
citations
#287

ShadowKV: KV Cache in Shadows for High-Throughput Long-Context LLM Inference

Hanshi Sun, Li-Wen Chang, Wenlei Bao et al.

ICML 2025spotlightarXiv:2410.21465
65
citations
#288

GliDe with a CaPE: A Low-Hassle Method to Accelerate Speculative Decoding

Cunxiao Du, Jing Jiang, Xu Yuanchen et al.

ICML 2024arXiv:2402.02082
65
citations
#289

NExT: Teaching Large Language Models to Reason about Code Execution

Ansong Ni, Miltiadis Allamanis, Arman Cohan et al.

ICML 2024arXiv:2404.14662
65
citations
#290

Inductive Moment Matching

Linqi (Alex) Zhou, Stefano Ermon, Jiaming Song

ICML 2025oralarXiv:2503.07565
65
citations
#291

Flow Q-Learning

Seohong Park, Qiyang Li, Sergey Levine

ICML 2025arXiv:2502.02538
65
citations
#292

RE-Bench: Evaluating Frontier AI R&D Capabilities of Language Model Agents against Human Experts

Hjalmar Wijk, Tao Lin, Joel Becker et al.

ICML 2025spotlightarXiv:2411.15114
65
citations
#293

Dense Reward for Free in Reinforcement Learning from Human Feedback

Alexander Chan, Hao Sun, Samuel Holt et al.

ICML 2024arXiv:2402.00782
65
citations
#294

Covert Malicious Finetuning: Challenges in Safeguarding LLM Adaptation

Danny Halawi, Alexander Wei, Eric Wallace et al.

ICML 2024arXiv:2406.20053
65
citations
#295

Mixtures of Experts Unlock Parameter Scaling for Deep RL

Johan Obando Ceron, Ghada Sokar, Timon Willi et al.

ICML 2024spotlightarXiv:2402.08609
64
citations
#296

What needs to go right for an induction head? A mechanistic study of in-context learning circuits and their formation

Aaditya Singh, Ted Moskovitz, Feilx Hill et al.

ICML 2024spotlightarXiv:2404.07129
64
citations
#297

Prompt-tuning Latent Diffusion Models for Inverse Problems

Hyungjin Chung, Jong Chul YE, Peyman Milanfar et al.

ICML 2024arXiv:2310.01110
64
citations
#298

Fast Video Generation with Sliding Tile Attention

Peiyuan Zhang, Yongqi Chen, Runlong Su et al.

ICML 2025oralarXiv:2502.04507
64
citations
#299

Should we be going MAD? A Look at Multi-Agent Debate Strategies for LLMs

Andries Smit, Nathan Grinsztajn, Paul Duckworth et al.

ICML 2024arXiv:2311.17371
64
citations
#300

MMPareto: Boosting Multimodal Learning with Innocent Unimodal Assistance

Yake Wei, Di Hu

ICML 2024arXiv:2405.17730
64
citations
#301

An analytic theory of creativity in convolutional diffusion models

Mason Kamb, Surya Ganguli

ICML 2025oralarXiv:2412.20292
64
citations
#302

DoraemonGPT: Toward Understanding Dynamic Scenes with Large Language Models (Exemplified as A Video Agent)

Zongxin Yang, Guikun Chen, Xiaodi Li et al.

ICML 2024oralarXiv:2401.08392
64
citations
#303

KernelBench: Can LLMs Write Efficient GPU Kernels?

Anne Ouyang, Simon Guo, Simran Arora et al.

ICML 2025arXiv:2502.10517
63
citations
#304

MM-RLHF: The Next Step Forward in Multimodal LLM Alignment

Yi-Fan Zhang, Tao Yu, Haochen Tian et al.

ICML 2025arXiv:2502.10391
63
citations
#305

MATH-Perturb: Benchmarking LLMs' Math Reasoning Abilities against Hard Perturbations

Kaixuan Huang, Jiacheng Guo, Zihao Li et al.

ICML 2025arXiv:2502.06453
63
citations
#306

Transformers Implement Functional Gradient Descent to Learn Non-Linear Functions In Context

Xiang Cheng, Yuxin Chen, Suvrit Sra

ICML 2024arXiv:2312.06528
63
citations
#307

Differentially Private Synthetic Data via Foundation Model APIs 2: Text

Chulin Xie, Zinan Lin, Arturs Backurs et al.

ICML 2024spotlightarXiv:2403.01749
63
citations
#308

Position: Topological Deep Learning is the New Frontier for Relational Learning

Theodore Papamarkou, Tolga Birdal, Michael Bronstein et al.

ICML 2024arXiv:2402.08871
63
citations
#309

OptiMUS: Scalable Optimization Modeling with (MI)LP Solvers and Large Language Models

Ali AhmadiTeshnizi, Wenzhi Gao, Madeleine Udell

ICML 2024arXiv:2402.10172
62
citations
#310

Getting the most out of your tokenizer for pre-training and domain adaptation

Gautier Dagan, Gabriel Synnaeve, Baptiste Roziere

ICML 2024arXiv:2402.01035
62
citations
#311

Craftax: A Lightning-Fast Benchmark for Open-Ended Reinforcement Learning

Michael Matthews, Michael Beukman, Benjamin Ellis et al.

ICML 2024spotlightarXiv:2402.16801
62
citations
#312

Revitalizing Multivariate Time Series Forecasting: Learnable Decomposition with Inter-Series Dependencies and Intra-Series Variations Modeling

Guoqi Yu, Jing Zou, Xiaowei Hu et al.

ICML 2024arXiv:2402.12694
62
citations
#313

All-in-one simulation-based inference

Manuel Gloeckler, Michael Deistler, Christian Weilbach et al.

ICML 2024arXiv:2404.09636
62
citations
#314

Align Your Steps: Optimizing Sampling Schedules in Diffusion Models

Amirmojtaba Sabour, Sanja Fidler, Karsten Kreis

ICML 2024arXiv:2404.14507
61
citations
#315

Automatically Interpreting Millions of Features in Large Language Models

Gonçalo Paulo, Alex Mallen, Caden Juang et al.

ICML 2025arXiv:2410.13928
61
citations
#316

Zero-Shot ECG Classification with Multimodal Learning and Test-time Clinical Knowledge Enhancement

che liu, Zhongwei Wan, Cheng Ouyang et al.

ICML 2024arXiv:2403.06659
61
citations
#317

Federated Full-Parameter Tuning of Billion-Sized Language Models with Communication Cost under 18 Kilobytes

Zhen Qin, Daoyuan Chen, Bingchen Qian et al.

ICML 2024arXiv:2312.06353
61
citations
#318

Test-Time Model Adaptation with Only Forward Passes

Shuaicheng Niu, Chunyan Miao, Guohao Chen et al.

ICML 2024arXiv:2404.01650
60
citations
#319

Training-Free Long-Context Scaling of Large Language Models

Chenxin An, Fei Huang, Jun Zhang et al.

ICML 2024arXiv:2402.17463
60
citations
#320

Transformers, parallel computation, and logarithmic depth

Clayton Sanford, Daniel Hsu, Matus Telgarsky

ICML 2024spotlightarXiv:2402.09268
60
citations
#321

EE-LLM: Large-Scale Training and Inference of Early-Exit Large Language Models with 3D Parallelism

Yanxi Chen, Xuchen Pan, Yaliang Li et al.

ICML 2024arXiv:2312.04916
60
citations
#322

Position: Bayesian Deep Learning is Needed in the Age of Large-Scale AI

Theodore Papamarkou, Maria Skoularidou, Konstantina Palla et al.

ICML 2024arXiv:2402.00809
60
citations
#323

Position: The No Free Lunch Theorem, Kolmogorov Complexity, and the Role of Inductive Biases in Machine Learning

Micah Goldblum, Marc Finzi, Keefer Rowan et al.

ICML 2024spotlight
60
citations
#324

Parameter-Efficient Fine-Tuning with Discrete Fourier Transform

Ziqi Gao, Qichao Wang, Aochuan Chen et al.

ICML 2024arXiv:2405.03003
60
citations
#325

Learning to Plan & Reason for Evaluation with Thinking-LLM-as-a-Judge

Swarnadeep Saha, Xian Li, Marjan Ghazvininejad et al.

ICML 2025arXiv:2501.18099
60
citations
#326

AutoML-Agent: A Multi-Agent LLM Framework for Full-Pipeline AutoML

Patara Trirat, Wonyong Jeong, Sung Ju Hwang

ICML 2025arXiv:2410.02958
60
citations
#327

Feedback Loops With Language Models Drive In-Context Reward Hacking

Alexander Pan, Erik Jones, Meena Jagadeesan et al.

ICML 2024arXiv:2402.06627
60
citations
#328

Decoding-time Realignment of Language Models

Tianlin Liu, Shangmin Guo, Leonardo Martins Bianco et al.

ICML 2024spotlightarXiv:2402.02992
59
citations
#329

Normalizing Flows are Capable Generative Models

Shuangfei Zhai, Ruixiang Zhang, Preetum Nakkiran et al.

ICML 2025oralarXiv:2412.06329
59
citations
#330

Zero-Shot Unsupervised and Text-Based Audio Editing Using DDPM Inversion

Hila Manor, Tomer Michaeli

ICML 2024arXiv:2402.10009
59
citations
#331

InstructZero: Efficient Instruction Optimization for Black-Box Large Language Models

Lichang Chen, Jiuhai Chen, Tom Goldstein et al.

ICML 2024arXiv:2306.03082
59
citations
#332

MVMoE: Multi-Task Vehicle Routing Solver with Mixture-of-Experts

Jianan Zhou, Zhiguang Cao, Yaoxin Wu et al.

ICML 2024arXiv:2405.01029
59
citations
#333

Foundation Policies with Hilbert Representations

Seohong Park, Tobias Kreiman, Sergey Levine

ICML 2024oralarXiv:2402.15567
59
citations
#334

Learning Multi-Level Features with Matryoshka Sparse Autoencoders

Bart Bussmann, Noa Nabeshima, Adam Karvonen et al.

ICML 2025arXiv:2503.17547
58
citations
#335

Towards Modular LLMs by Building and Reusing a Library of LoRAs

Oleksiy Ostapenko, Zhan Su, Edoardo Ponti et al.

ICML 2024arXiv:2405.11157
58
citations
#336

CLLMs: Consistency Large Language Models

Siqi Kou, Lanxiang Hu, Zhezhi He et al.

ICML 2024arXiv:2403.00835
58
citations
#337

SAEBench: A Comprehensive Benchmark for Sparse Autoencoders in Language Model Interpretability

Adam Karvonen, Can Rager, Johnny Lin et al.

ICML 2025arXiv:2503.09532
58
citations
#338

Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning

Zhiheng Xi, Wenxiang Chen, Boyang Hong et al.

ICML 2024arXiv:2402.05808
58
citations
#339

On the Origins of Linear Representations in Large Language Models

Yibo Jiang, Goutham Rajendran, Pradeep Ravikumar et al.

ICML 2024arXiv:2403.03867
58
citations
#340

CogBench: a large language model walks into a psychology lab

Julian Coda-Forno, Marcel Binz, Jane Wang et al.

ICML 2024oralarXiv:2402.18225
57
citations
#341

Learning to Route Among Specialized Experts for Zero-Shot Generalization

Mohammed Muqeeth, Haokun Liu, Yufan Liu et al.

ICML 2024arXiv:2402.05859
57
citations
#342

NoLiMa: Long-Context Evaluation Beyond Literal Matching

Ali Modarressi, Hanieh Deilamsalehy, Franck Dernoncourt et al.

ICML 2025arXiv:2502.05167
57
citations
#343

VisionTS: Visual Masked Autoencoders Are Free-Lunch Zero-Shot Time Series Forecasters

Mouxiang Chen, Lefei Shen, Zhuo Li et al.

ICML 2025arXiv:2408.17253
56
citations
#344

VinePPO: Refining Credit Assignment in RL Training of LLMs

Amirhossein Kazemnejad, Milad Aghajohari, Eva Portelance et al.

ICML 2025arXiv:2410.01679
56
citations
#345

SAMformer: Unlocking the Potential of Transformers in Time Series Forecasting with Sharpness-Aware Minimization and Channel-Wise Attention

Romain Ilbert, Ambroise Odonnat, Vasilii Feofanov et al.

ICML 2024arXiv:2402.10198
56
citations
#346

Position: What Can Large Language Models Tell Us about Time Series Analysis

Ming Jin, Yi-Fan Zhang, Wei Chen et al.

ICML 2024arXiv:2402.02713
56
citations
#347

CompeteAI: Understanding the Competition Dynamics of Large Language Model-based Agents

Qinlin Zhao, Jindong Wang, Yixuan Zhang et al.

ICML 2024arXiv:2310.17512
56
citations
#348

Are Sparse Autoencoders Useful? A Case Study in Sparse Probing

Subhash Kantamneni, Josh Engels, Senthooran Rajamanoharan et al.

ICML 2025arXiv:2502.16681
56
citations
#349

Probabilistic Inference in Language Models via Twisted Sequential Monte Carlo

Stephen Zhao, Rob Brekelmans, Alireza Makhzani et al.

ICML 2024arXiv:2404.17546
56
citations
#350

Neural Operators with Localized Integral and Differential Kernels

Miguel Liu-Schiaffini, Julius Berner, Boris Bonev et al.

ICML 2024arXiv:2402.16845
56
citations
#351

Instruction Tuning for Secure Code Generation

Jingxuan He, Mark Vero, Gabriela Krasnopolska et al.

ICML 2024arXiv:2402.09497
56
citations
#352

Dynamic Evaluation of Large Language Models by Meta Probing Agents

Kaijie Zhu, Jindong Wang, Qinlin Zhao et al.

ICML 2024arXiv:2402.14865
55
citations
#353

Auto-Regressive Next-Token Predictors are Universal Learners

Eran Malach

ICML 2024arXiv:2309.06979
55
citations
#354

Adaptive Text Watermark for Large Language Models

Yepeng Liu, Yuheng Bu

ICML 2024arXiv:2401.13927
55
citations
#355

How Universal Polynomial Bases Enhance Spectral Graph Neural Networks: Heterophily, Over-smoothing, and Over-squashing

Keke Huang, Yu Guang Wang, Ming Li et al.

ICML 2024arXiv:2405.12474
55
citations
#356

Differentially Private Bias-Term Fine-tuning of Foundation Models

Zhiqi Bu, Yu-Xiang Wang, Sheng Zha et al.

ICML 2024arXiv:2210.00036
55
citations
#357

Swallowing the Bitter Pill: Simplified Scalable Conformer Generation

Yuyang Wang, Ahmed Elhag, Navdeep Jaitly et al.

ICML 2024arXiv:2311.17932
54
citations
#358

Tag-LLM: Repurposing General-Purpose LLMs for Specialized Domains

Junhong Shen, Neil Tenenholtz, James Hall et al.

ICML 2024arXiv:2402.05140
54
citations
#359

Pruner-Zero: Evolving Symbolic Pruning Metric From Scratch for Large Language Models

Peijie Dong, Lujun Li, Zhenheng Tang et al.

ICML 2024arXiv:2406.02924
54
citations
#360

DéjàVu: KV-cache Streaming for Fast, Fault-tolerant Generative LLM Serving

Foteini Strati, Sara McAllister, Amar Phanishayee et al.

ICML 2024arXiv:2403.01876
53
citations
#361

Organize the Web: Constructing Domains Enhances Pre-Training Data Curation

Alexander Wettig, Kyle Lo, Sewon Min et al.

ICML 2025arXiv:2502.10341
53
citations
#362

Mechanistic Design and Scaling of Hybrid Architectures

Michael Poli, Armin Thomas, Eric Nguyen et al.

ICML 2024arXiv:2403.17844
53
citations
#363

Long-Term TalkingFace Generation via Motion-Prior Conditional Diffusion Model

SHEN FEI, Cong Wang, Junyao Gao et al.

ICML 2025oralarXiv:2502.09533
53
citations
#364

A Diffusion Model Framework for Unsupervised Neural Combinatorial Optimization

Sebastian Sanokowski, Sepp Hochreiter, Sebastian Lehner

ICML 2024arXiv:2406.01661
53
citations
#365

AquaLoRA: Toward White-box Protection for Customized Stable Diffusion Models via Watermark LoRA

Weitao Feng, Wenbo Zhou, Jiyan He et al.

ICML 2024arXiv:2405.11135
53
citations
#366

Particle Denoising Diffusion Sampler

Angus Phillips, Hai-Dang Dau, Michael Hutchinson et al.

ICML 2024arXiv:2402.06320
53
citations
#367

MolCRAFT: Structure-Based Drug Design in Continuous Parameter Space

Yanru Qu, Keyue Qiu, Yuxuan Song et al.

ICML 2024arXiv:2404.12141
52
citations
#368

Token Assorted: Mixing Latent and Text Tokens for Improved Language Model Reasoning

Andy (DiJia) Su, Hanlin Zhu, Yingchen Xu et al.

ICML 2025arXiv:2502.03275
52
citations
#369

Premise Order Matters in Reasoning with Large Language Models

Xinyun Chen, Ryan Chi, Xuezhi Wang et al.

ICML 2024arXiv:2402.08939
52
citations
#370

Sequential Neural Score Estimation: Likelihood-Free Inference with Conditional Score Based Diffusion Models

Louis Sharrock, Jack Simons, Song Liu et al.

ICML 2024spotlightarXiv:2210.04872
52
citations
#371

Cooperative Graph Neural Networks

Ben Finkelshtein, Xingyue Huang, Michael Bronstein et al.

ICML 2024arXiv:2310.01267
52
citations
#372

Masked Autoencoders Are Effective Tokenizers for Diffusion Models

Hao Chen, Yujin Han, Fangyi Chen et al.

ICML 2025spotlightarXiv:2502.03444
52
citations
#373

GeminiFusion: Efficient Pixel-wise Multimodal Fusion for Vision Transformer

Ding Jia, Jianyuan Guo, Kai Han et al.

ICML 2024arXiv:2406.01210
51
citations
#374

G-Designer: Architecting Multi-agent Communication Topologies via Graph Neural Networks

Guibin Zhang, Yanwei Yue, Xiangguo Sun et al.

ICML 2025spotlightarXiv:2410.11782
51
citations
#375

EQ-VAE: Equivariance Regularized Latent Space for Improved Generative Image Modeling

Theodoros Kouzelis, Ioannis Kakogeorgiou, Spyros Gidaris et al.

ICML 2025arXiv:2502.09509
51
citations
#376

Scaling Exponents Across Parameterizations and Optimizers

Katie Everett, Lechao Xiao, Mitchell Wortsman et al.

ICML 2024arXiv:2407.05872
51
citations
#377

NETS: A Non-equilibrium Transport Sampler

Michael Albergo, Eric Vanden-Eijnden

ICML 2025arXiv:2410.02711
51
citations
#378

Curated LLM: Synergy of LLMs and Data Curation for tabular augmentation in low-data regimes

Nabeel Seedat, Nicolas Huynh, Boris van Breugel et al.

ICML 2024arXiv:2312.12112
51
citations
#379

FlexTok: Resampling Images into 1D Token Sequences of Flexible Length

Roman Bachmann, Jesse Allardice, David Mizrahi et al.

ICML 2025arXiv:2502.13967
51
citations
#380

SelfIE: Self-Interpretation of Large Language Model Embeddings

Haozhe Chen, Carl Vondrick, Chengzhi Mao

ICML 2024arXiv:2403.10949
51
citations
#381

Critical Tokens Matter: Token-Level Contrastive Estimation Enhances LLM’s Reasoning Capability

Zicheng Lin, Tian Liang, Jiahao Xu et al.

ICML 2025arXiv:2411.19943
51
citations
#382

All-atom Diffusion Transformers: Unified generative modelling of molecules and materials

Chaitanya Joshi, Xiang Fu, Yi-Lun Liao et al.

ICML 2025arXiv:2503.03965
50
citations
#383

Monte Carlo Tree Search for Comprehensive Exploration in LLM-Based Automatic Heuristic Design

Zhi Zheng, Zhuoliang Xie, Zhenkun Wang et al.

ICML 2025arXiv:2501.08603
50
citations
#384

Protein Conformation Generation via Force-Guided SE(3) Diffusion Models

YAN WANG, Lihao Wang, Yuning Shen et al.

ICML 2024arXiv:2403.14088
50
citations
#385

ReFocus: Visual Editing as a Chain of Thought for Structured Image Understanding

Xingyu Fu, Minqian Liu, Zhengyuan Yang et al.

ICML 2025arXiv:2501.05452
49
citations
#386

Causal Representation Learning from Multiple Distributions: A General Setting

Kun Zhang, Shaoan Xie, Ignavier Ng et al.

ICML 2024oralarXiv:2402.05052
49
citations
#387

EraseAnything: Enabling Concept Erasure in Rectified Flow Transformers

Daiheng Gao, Shilin Lu, Wenbo Zhou et al.

ICML 2025arXiv:2412.20413
49
citations
#388

Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression

Junyuan Hong, Jinhao Duan, Chenhui Zhang et al.

ICML 2024arXiv:2403.15447
49
citations
#389

Why Larger Language Models Do In-context Learning Differently?

Zhenmei Shi, Junyi Wei, Zhuoyan Xu et al.

ICML 2024arXiv:2405.19592
49
citations
#390

RoSA: Accurate Parameter-Efficient Fine-Tuning via Robust Adaptation

Mahdi Nikdan, Soroush Tabesh, Elvir Crnčević et al.

ICML 2024arXiv:2401.04679
48
citations
#391

Iterative Data Smoothing: Mitigating Reward Overfitting and Overoptimization in RLHF

Banghua Zhu, Michael Jordan, Jiantao Jiao

ICML 2024arXiv:2401.16335
48
citations
#392

Self-Alignment of Large Language Models via Monopolylogue-based Social Scene Simulation

Xianghe Pang, shuo tang, Rui Ye et al.

ICML 2024spotlightarXiv:2402.05699
48
citations
#393

Model Tailor: Mitigating Catastrophic Forgetting in Multi-modal Large Language Models

Didi Zhu, Zhongyi Sun, Zexi Li et al.

ICML 2024arXiv:2402.12048
48
citations
#394

Linguistic Calibration of Long-Form Generations

Neil Band, Xuechen Li, Tengyu Ma et al.

ICML 2024arXiv:2404.00474
48
citations
#395

Antidote: Post-fine-tuning Safety Alignment for Large Language Models against Harmful Fine-tuning Attack

Tiansheng Huang, Gautam Bhattacharya, Pratik Joshi et al.

ICML 2025
48
citations
#396

STAIR: Improving Safety Alignment with Introspective Reasoning

Yichi Zhang, Siyuan Zhang, Yao Huang et al.

ICML 2025oralarXiv:2502.02384
48
citations
#397

FlowAR: Scale-wise Autoregressive Image Generation Meets Flow Matching

Sucheng Ren, Qihang Yu, Ju He et al.

ICML 2025arXiv:2412.15205
47
citations
#398

Variational Learning is Effective for Large Deep Networks

Yuesong Shen, Nico Daheim, Bai Cong et al.

ICML 2024spotlightarXiv:2402.17641
47
citations
#399

FlipAttack: Jailbreak LLMs via Flipping

Yue Liu, Xiaoxin He, Miao Xiong et al.

ICML 2025arXiv:2410.02832
47
citations
#400

Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration

Zhongzhi Yu, Zheng Wang, Yonggan Fu et al.

ICML 2024arXiv:2406.15765
47
citations