Most Cited ICLR "adaptive constrained equivariance" Papers

6,124 papers found • Page 3 of 31

#401

Synapse: Trajectory-as-Exemplar Prompting with Memory for Computer Control

Longtao Zheng, Rundong Wang, Xinrun Wang et al.

ICLR 2024arXiv:2306.07863
110
citations
#402

Autoregressive Video Generation without Vector Quantization

Haoge Deng, Ting Pan, Haiwen Diao et al.

ICLR 2025oralarXiv:2412.14169
110
citations
#403

Unpaired Image-to-Image Translation via Neural Schrödinger Bridge

Beomsu Kim, Gihyun Kwon, Kwanyoung Kim et al.

ICLR 2024arXiv:2305.15086
109
citations
#404

On the self-verification limitations of large language models on reasoning and planning tasks

Kaya Stechly, Karthik Valmeekam, Subbarao Kambhampati

ICLR 2025arXiv:2402.08115
109
citations
#405

#InsTag: Instruction Tagging for Analyzing Supervised Fine-tuning of Large Language Models

Keming Lu, Hongyi Yuan, Zheng Yuan et al.

ICLR 2024arXiv:2308.07074
109
citations
#406

SV4D: Dynamic 3D Content Generation with Multi-Frame and Multi-View Consistency

Yiming Xie, Chun-Han Yao, Vikram Voleti et al.

ICLR 2025oralarXiv:2407.17470
108
citations
#407

Retroformer: Retrospective Large Language Agents with Policy Gradient Optimization

Weiran Yao, Shelby Heinecke, Juan Carlos Niebles et al.

ICLR 2024spotlightarXiv:2308.02151
108
citations
#408

A Paradigm Shift in Machine Translation: Boosting Translation Performance of Large Language Models

Haoran Xu, Young Jin Kim, Amr Mohamed Nabil Aly Aly Sharaf et al.

ICLR 2024arXiv:2309.11674
108
citations
#409

SWE-bench Multimodal: Do AI Systems Generalize to Visual Software Domains?

John Yang, Carlos E Jimenez, Alex Zhang et al.

ICLR 2025arXiv:2410.03859
108
citations
#410

OmniRe: Omni Urban Scene Reconstruction

Ziyu Chen, Jiawei Yang, Jiahui Huang et al.

ICLR 2025arXiv:2408.16760
107
citations
#411

Agent S: An Open Agentic Framework that Uses Computers Like a Human

Saaket Agashe, Jiuzhou Han, Shuyu Gan et al.

ICLR 2025arXiv:2410.08164
107
citations
#412

Sudden Drops in the Loss: Syntax Acquisition, Phase Transitions, and Simplicity Bias in MLMs

Angelica Chen, Ravid Shwartz-Ziv, Kyunghyun Cho et al.

ICLR 2024spotlightarXiv:2309.07311
107
citations
#413

Fine-Tuned Language Models Generate Stable Inorganic Materials as Text

Nate Gruver, Anuroop Sriram, Andrea Madotto et al.

ICLR 2024arXiv:2402.04379
107
citations
#414

Tensor Trust: Interpretable Prompt Injection Attacks from an Online Game

Sam Toyer, Olivia Watkins, Ethan Mendes et al.

ICLR 2024spotlightarXiv:2311.01011
106
citations
#415

Semantic Image Inversion and Editing using Rectified Stochastic Differential Equations

Litu Rout, Yujia Chen, Nataniel Ruiz et al.

ICLR 2025arXiv:2410.10792
106
citations
#416

Consistent4D: Consistent 360° Dynamic Object Generation from Monocular Video

Yanqin Jiang, Li Zhang, Jin Gao et al.

ICLR 2024oralarXiv:2311.02848
106
citations
#417

MS-Diffusion: Multi-subject Zero-shot Image Personalization with Layout Guidance

Xierui Wang, Siming Fu, Qihan Huang et al.

ICLR 2025arXiv:2406.07209
106
citations
#418

VideoPhy: Evaluating Physical Commonsense for Video Generation

Hritik Bansal, Zongyu Lin, Tianyi Xie et al.

ICLR 2025arXiv:2406.03520
106
citations
#419

AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark

Wenhao Chai, Enxin Song, Yilun Du et al.

ICLR 2025oralarXiv:2410.03051
105
citations
#420

ToolChain*: Efficient Action Space Navigation in Large Language Models with A* Search

Yuchen Zhuang, Xiang Chen, Tong Yu et al.

ICLR 2024arXiv:2310.13227
105
citations
#421

CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets

Lifan Yuan, Yangyi Chen, Xingyao Wang et al.

ICLR 2024arXiv:2309.17428
105
citations
#422

RegMix: Data Mixture as Regression for Language Model Pre-training

Qian Liu, Xiaosen Zheng, Niklas Muennighoff et al.

ICLR 2025arXiv:2407.01492
105
citations
#423

HIFA: High-fidelity Text-to-3D Generation with Advanced Diffusion Guidance

Junzhe Zhu, Peiye Zhuang, Sanmi Koyejo

ICLR 2024arXiv:2305.18766
105
citations
#424

Large Language Models as Generalizable Policies for Embodied Tasks

Andrew Szot, Max Schwarzer, Harsh Agrawal et al.

ICLR 2024arXiv:2310.17722
105
citations
#425

Conformal Language Modeling

Victor Quach, Adam Fisch, Tal Schuster et al.

ICLR 2024arXiv:2306.10193
105
citations
#426

Knowledge Fusion of Large Language Models

Fanqi Wan, Xinting Huang, Deng Cai et al.

ICLR 2024arXiv:2401.10491
104
citations
#427

Long-Context LLMs Meet RAG: Overcoming Challenges for Long Inputs in RAG

Bowen Jin, Jinsung Yoon, Jiawei Han et al.

ICLR 2025arXiv:2410.05983
104
citations
#428

ReLU Strikes Back: Exploiting Activation Sparsity in Large Language Models

Iman Mirzadeh, Keivan Alizadeh-Vahid, Sachin Mehta et al.

ICLR 2024arXiv:2310.04564
104
citations
#429

PoSE: Efficient Context Window Extension of LLMs via Positional Skip-wise Training

Dawei Zhu, Nan Yang, Liang Wang et al.

ICLR 2024arXiv:2309.10400
103
citations
#430

LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs

Yushi Bai, Jiajie Zhang, Xin Lv et al.

ICLR 2025arXiv:2408.07055
103
citations
#431

BadEdit: Backdooring Large Language Models by Model Editing

Yanzhou Li, Tianlin Li, Kangjie Chen et al.

ICLR 2024arXiv:2403.13355
102
citations
#432

Vision-by-Language for Training-Free Compositional Image Retrieval

Shyamgopal Karthik, Karsten Roth, Massimiliano Mancini et al.

ICLR 2024arXiv:2310.09291
102
citations
#433

Feast Your Eyes: Mixture-of-Resolution Adaptation for Multimodal Large Language Models

Gen Luo, Yiyi Zhou, Yuxin Zhang et al.

ICLR 2025arXiv:2403.03003
102
citations
#434

VDT: General-purpose Video Diffusion Transformers via Mask Modeling

Haoyu Lu, Guoxing Yang, Nanyi Fei et al.

ICLR 2024oralarXiv:2305.13311
102
citations
#435

UniversalNER: Targeted Distillation from Large Language Models for Open Named Entity Recognition

Wenxuan Zhou, Sheng Zhang, Yu Gu et al.

ICLR 2024arXiv:2308.03279
102
citations
#436

Universal Humanoid Motion Representations for Physics-Based Control

Zhengyi Luo, Jinkun Cao, Josh Merel et al.

ICLR 2024spotlightarXiv:2310.04582
102
citations
#437

MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers

Yiwen Chen, Tong He, Di Huang et al.

ICLR 2025arXiv:2406.10163
102
citations
#438

Not All Language Model Features Are One-Dimensionally Linear

Josh Engels, Eric Michaud, Isaac Liao et al.

ICLR 2025arXiv:2405.14860
101
citations
#439

Unified Human-Scene Interaction via Prompted Chain-of-Contacts

Zeqi Xiao, Tai Wang, Jingbo Wang et al.

ICLR 2024spotlightarXiv:2309.07918
101
citations
#440

Tag2Text: Guiding Vision-Language Model via Image Tagging

Xinyu Huang, Youcai Zhang, Jinyu Ma et al.

ICLR 2024arXiv:2303.05657
101
citations
#441

The Expressive Power of Low-Rank Adaptation

Yuchen Zeng, Kangwook Lee

ICLR 2024arXiv:2310.17513
101
citations
#442

Language Model Cascades: Token-Level Uncertainty And Beyond

Neha Gupta, Harikrishna Narasimhan, Wittawat Jitkrittum et al.

ICLR 2024arXiv:2404.10136
101
citations
#443

Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process

Tian Ye, Zicheng Xu, Yuanzhi Li et al.

ICLR 2025arXiv:2407.20311
100
citations
#444

HART: Efficient Visual Generation with Hybrid Autoregressive Transformer

Haotian Tang, Yecheng Wu, Shang Yang et al.

ICLR 2025arXiv:2410.10812
100
citations
#445

Circuit Component Reuse Across Tasks in Transformer Language Models

Jack Merullo, Carsten Eickhoff, Ellie Pavlick

ICLR 2024spotlightarXiv:2310.08744
99
citations
#446

Consistency Models Made Easy

Zhengyang Geng, Ashwini Pokle, Weijian Luo et al.

ICLR 2025arXiv:2406.14548
99
citations
#447

DataInf: Efficiently Estimating Data Influence in LoRA-tuned LLMs and Diffusion Models

Yongchan Kwon, Eric Wu, Kevin Wu et al.

ICLR 2024arXiv:2310.00902
99
citations
#448

Towards image compression with perfect realism at ultra-low bitrates

Marlene Careil, Matthew J Muckley, Jakob Verbeek et al.

ICLR 2024arXiv:2310.10325
99
citations
#449

Decoding Natural Images from EEG for Object Recognition

Yonghao Song, Bingchuan Liu, Xiang Li et al.

ICLR 2024oralarXiv:2308.13234
99
citations
#450

Distributional Preference Learning: Understanding and Accounting for Hidden Context in RLHF

Anand Siththaranjan, Cassidy Laidlaw, Dylan Hadfield-Menell

ICLR 2024arXiv:2312.08358
99
citations
#451

Fine-Tuning Enhances Existing Mechanisms: A Case Study on Entity Tracking

Nikhil Prakash, Tamar Shaham, Tal Haklay et al.

ICLR 2024arXiv:2402.14811
99
citations
#452

Parameter-Efficient Orthogonal Finetuning via Butterfly Factorization

Weiyang Liu, Zeju Qiu, Yao Feng et al.

ICLR 2024arXiv:2311.06243
98
citations
#453

SVDQuant: Absorbing Outliers by Low-Rank Component for 4-Bit Diffusion Models

Muyang Li, Yujun Lin, Zhekai Zhang et al.

ICLR 2025arXiv:2411.05007
98
citations
#454

Rethinking Model Ensemble in Transfer-based Adversarial Attacks

Huanran Chen, Yichi Zhang, Yinpeng Dong et al.

ICLR 2024arXiv:2303.09105
98
citations
#455

Kosmos-G: Generating Images in Context with Multimodal Large Language Models

Xichen Pan, Li Dong, Shaohan Huang et al.

ICLR 2024arXiv:2310.02992
98
citations
#456

DFormer: Rethinking RGBD Representation Learning for Semantic Segmentation

Bowen Yin, Xuying Zhang, Zhong-Yu Li et al.

ICLR 2024arXiv:2309.09668
98
citations
#457

HyperAttention: Long-context Attention in Near-Linear Time

Insu Han, Rajesh Jayaram, Amin Karbasi et al.

ICLR 2024arXiv:2310.05869
98
citations
#458

Adam-mini: Use Fewer Learning Rates To Gain More

Yushun Zhang, Congliang Chen, Ziniu Li et al.

ICLR 2025arXiv:2406.16793
98
citations
#459

When Attention Sink Emerges in Language Models: An Empirical View

Xiangming Gu, Tianyu Pang, Chao Du et al.

ICLR 2025arXiv:2410.10781
98
citations
#460

Mechanistically analyzing the effects of fine-tuning on procedurally defined tasks

Samyak Jain, Robert Kirk, Ekdeep Singh Lubana et al.

ICLR 2024arXiv:2311.12786
97
citations
#461

A Semantic Invariant Robust Watermark for Large Language Models

Aiwei Liu, Leyi Pan, Xuming Hu et al.

ICLR 2024arXiv:2310.06356
96
citations
#462

ARGS: Alignment as Reward-Guided Search

Maxim Khanov, Jirayu Burapacheep, Yixuan Li

ICLR 2024arXiv:2402.01694
96
citations
#463

CADS: Unleashing the Diversity of Diffusion Models through Condition-Annealed Sampling

Seyedmorteza Sadat, Jakob Buhmann, Derek Bradley et al.

ICLR 2024spotlightarXiv:2310.17347
96
citations
#464

Noise-free Score Distillation

Oren Katzir, Or Patashnik, Daniel Cohen-Or et al.

ICLR 2024arXiv:2310.17590
96
citations
#465

LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias

Haian Jin, Hanwen Jiang, Hao Tan et al.

ICLR 2025arXiv:2410.17242
96
citations
#466

CBraMod: A Criss-Cross Brain Foundation Model for EEG Decoding

Jiquan Wang, Sha Zhao, Zhiling Luo et al.

ICLR 2025oralarXiv:2412.07236
95
citations
#467

At Which Training Stage Does Code Data Help LLMs Reasoning?

ma yingwei, Yue Liu, Yue Yu et al.

ICLR 2024spotlightarXiv:2309.16298
95
citations
#468

Deconstructing Denoising Diffusion Models for Self-Supervised Learning

Xinlei Chen, Zhuang Liu, Saining Xie et al.

ICLR 2025arXiv:2401.14404
95
citations
#469

SageAttention: Accurate 8-Bit Attention for Plug-and-play Inference Acceleration

Jintao Zhang, Jia wei, Pengle Zhang et al.

ICLR 2025arXiv:2410.02367
95
citations
#470

Loopy: Taming Audio-Driven Portrait Avatar with Long-Term Motion Dependency

Jianwen Jiang, Chao Liang, Jiaqi Yang et al.

ICLR 2025oralarXiv:2409.02634
95
citations
#471

Label-free Node Classification on Graphs with Large Language Models (LLMs)

Zhikai Chen, Haitao Mao, Hongzhi Wen et al.

ICLR 2024arXiv:2310.04668
95
citations
#472

DreamBench++: A Human-Aligned Benchmark for Personalized Image Generation

Yuang Peng, Yuxin Cui, Haomiao Tang et al.

ICLR 2025arXiv:2406.16855
95
citations
#473

MiniLLM: Knowledge Distillation of Large Language Models

Yuxian Gu, Li Dong, Furu Wei et al.

ICLR 2024arXiv:2306.08543
95
citations
#474

Privacy-Preserving In-Context Learning with Differentially Private Few-Shot Generation

Xinyu Tang, Richard Shin, Huseyin Inan et al.

ICLR 2024arXiv:2309.11765
94
citations
#475

Turning Up the Heat: Min-p Sampling for Creative and Coherent LLM Outputs

Minh Nguyen, Andrew Baker, Clement Neo et al.

ICLR 2025arXiv:2407.01082
94
citations
#476

ColPali: Efficient Document Retrieval with Vision Language Models

Manuel Faysse, Hugues Sibille, Tony Wu et al.

ICLR 2025arXiv:2407.01449
94
citations
#477

AdaIR: Adaptive All-in-One Image Restoration via Frequency Mining and Modulation

Yuning Cui, Syed Waqas Zamir, Salman Khan et al.

ICLR 2025arXiv:2403.14614
94
citations
#478

Copilot4D: Learning Unsupervised World Models for Autonomous Driving via Discrete Diffusion

Lunjun Zhang, Yuwen Xiong, Ze Yang et al.

ICLR 2024arXiv:2311.01017
94
citations
#479

FeatUp: A Model-Agnostic Framework for Features at Any Resolution

Stephanie Fu, Mark Hamilton, Laura E. Brandt et al.

ICLR 2024arXiv:2403.10516
93
citations
#480

Improved sampling via learned diffusions

Lorenz Richter, Julius Berner

ICLR 2024arXiv:2307.01198
93
citations
#481

TimeMixer++: A General Time Series Pattern Machine for Universal Predictive Analysis

Shiyu Wang, Jiawei LI, Xiaoming Shi et al.

ICLR 2025oralarXiv:2410.16032
93
citations
#482

Unbiased Watermark for Large Language Models

Zhengmian Hu, Lichang Chen, Xidong Wu et al.

ICLR 2024spotlightarXiv:2310.10669
93
citations
#483

Consistency-guided Prompt Learning for Vision-Language Models

Shuvendu Roy, Ali Etemad

ICLR 2024arXiv:2306.01195
93
citations
#484

Brain decoding: toward real-time reconstruction of visual perception

Yohann Benchetrit, Hubert Banville, Jean-Remi King

ICLR 2024oralarXiv:2310.19812
93
citations
#485

Predictive Inverse Dynamics Models are Scalable Learners for Robotic Manipulation

Yang Tian, Sizhe Yang, Jia Zeng et al.

ICLR 2025arXiv:2412.15109
93
citations
#486

An Extensible Framework for Open Heterogeneous Collaborative Perception

Yifan Lu, Yue Hu, Yiqi Zhong et al.

ICLR 2024arXiv:2401.13964
92
citations
#487

Bayesian Low-rank Adaptation for Large Language Models

Adam Yang, Maxime Robeyns, Xi Wang et al.

ICLR 2024arXiv:2308.13111
92
citations
#488

Kolmogorov-Arnold Transformer

Xingyi Yang, Xinchao Wang

ICLR 2025arXiv:2409.10594
92
citations
#489

OmniEdit: Building Image Editing Generalist Models Through Specialist Supervision

Cong Wei, Zheyang Xiong, Weiming Ren et al.

ICLR 2025arXiv:2411.07199
91
citations
#490

DynaMath: A Dynamic Visual Benchmark for Evaluating Mathematical Reasoning Robustness of Vision Language Models

Chengke Zou, Xingang Guo, Rui Yang et al.

ICLR 2025arXiv:2411.00836
91
citations
#491

Learning Delays in Spiking Neural Networks using Dilated Convolutions with Learnable Spacings

Ilyass Hammouamri, Ismail Khalfaoui Hassani, Timothée Masquelier

ICLR 2024oralarXiv:2306.17670
91
citations
#492

The Hedgehog & the Porcupine: Expressive Linear Attentions with Softmax Mimicry

Michael Zhang, Kush Bhatia, Hermann Kumbong et al.

ICLR 2024arXiv:2402.04347
91
citations
#493

Motif: Intrinsic Motivation from Artificial Intelligence Feedback

Martin Klissarov, Pierluca D'Oro, Shagun Sodhani et al.

ICLR 2024arXiv:2310.00166
91
citations
#494

GeoLLM: Extracting Geospatial Knowledge from Large Language Models

Rohin Manvi, Samar Khanna, Gengchen Mai et al.

ICLR 2024arXiv:2310.06213
91
citations
#495

CrossQ: Batch Normalization in Deep Reinforcement Learning for Greater Sample Efficiency and Simplicity

Aditya Bhatt, Daniel Palenicek, Boris Belousov et al.

ICLR 2024spotlightarXiv:1902.05605
91
citations
#496

Training Socially Aligned Language Models on Simulated Social Interactions

Ruibo Liu, Ruixin Yang, Chenyan Jia et al.

ICLR 2024arXiv:2305.16960
91
citations
#497

Turning large language models into cognitive models

Marcel Binz, Eric Schulz

ICLR 2024arXiv:2306.03917
90
citations
#498

Escape Sky-high Cost: Early-stopping Self-Consistency for Multi-step Reasoning

Yiwei Li, Peiwen Yuan, Shaoxiong Feng et al.

ICLR 2024arXiv:2401.10480
90
citations
#499

OGBench: Benchmarking Offline Goal-Conditioned RL

Seohong Park, Kevin Frans, Benjamin Eysenbach et al.

ICLR 2025arXiv:2410.20092
90
citations
#500

Real3D-Portrait: One-shot Realistic 3D Talking Portrait Synthesis

Zhenhui Ye, Tianyun Zhong, Yi Ren et al.

ICLR 2024spotlightarXiv:2401.08503
90
citations
#501

Fine-tuning Multimodal LLMs to Follow Zero-shot Demonstrative Instructions

Juncheng Li, Kaihang Pan, Zhiqi Ge et al.

ICLR 2024spotlightarXiv:2308.04152
90
citations
#502

Lemur: Harmonizing Natural Language and Code for Language Agents

Yiheng Xu, Hongjin SU, Chen Xing et al.

ICLR 2024spotlightarXiv:2310.06830
89
citations
#503

Making Text Embedders Few-Shot Learners

Chaofan Li, Minghao Qin, Shitao Xiao et al.

ICLR 2025arXiv:2409.15700
89
citations
#504

How Many Pretraining Tasks Are Needed for In-Context Learning of Linear Regression?

Jingfeng Wu, Difan Zou, Zixiang Chen et al.

ICLR 2024spotlightarXiv:2310.08391
89
citations
#505

SILO Language Models: Isolating Legal Risk In a Nonparametric Datastore

Sewon Min, Suchin Gururangan, Eric Wallace et al.

ICLR 2024spotlightarXiv:2308.04430
88
citations
#506

Unlocking Guidance for Discrete State-Space Diffusion and Flow Models

Hunter Nisonoff, Junhao Xiong, Stephan Allenspach et al.

ICLR 2025arXiv:2406.01572
88
citations
#507

LiveBench: A Challenging, Contamination-Limited LLM Benchmark

Colin White, Samuel Dooley, Manley Roberts et al.

ICLR 2025arXiv:2406.19314
88
citations
#508

Spoken Question Answering and Speech Continuation Using Spectrogram-Powered LLM

Eliya Nachmani, Alon Levkovitch, Roy Hirsch et al.

ICLR 2024arXiv:2305.15255
88
citations
#509

MLLMs Know Where to Look: Training-free Perception of Small Visual Details with Multimodal LLMs

jiarui zhang, Mahyar Khayatkhoei, Prateek Chhikara et al.

ICLR 2025arXiv:2502.17422
88
citations
#510

KoLA: Carefully Benchmarking World Knowledge of Large Language Models

Jifan Yu, Xiaozhi Wang, Shangqing Tu et al.

ICLR 2024arXiv:2306.09296
88
citations
#511

GraphRouter: A Graph-based Router for LLM Selections

Tao Feng, Yanzhen Shen, Jiaxuan You

ICLR 2025arXiv:2410.03834
87
citations
#512

Entropy is not Enough for Test-Time Adaptation: From the Perspective of Disentangled Factors

Jonghyun Lee, Dahuin Jung, Saehyung Lee et al.

ICLR 2024spotlightarXiv:2403.07366
87
citations
#513

Draw-and-Understand: Leveraging Visual Prompts to Enable MLLMs to Comprehend What You Want

Weifeng Lin, Xinyu Wei, Ruichuan An et al.

ICLR 2025arXiv:2403.20271
87
citations
#514

Finetuning Text-to-Image Diffusion Models for Fairness

Xudong Shen, Chao Du, Tianyu Pang et al.

ICLR 2024arXiv:2311.07604
87
citations
#515

Detecting, Explaining, and Mitigating Memorization in Diffusion Models

Yuxin Wen, Yuchen Liu, Chen Chen et al.

ICLR 2024arXiv:2407.21720
87
citations
#516

Batch Calibration: Rethinking Calibration for In-Context Learning and Prompt Engineering

Han Zhou, Xingchen Wan, Lev Proleev et al.

ICLR 2024arXiv:2309.17249
87
citations
#517

Vision-LSTM: xLSTM as Generic Vision Backbone

Benedikt Alkin, Maximilian Beck, Korbinian Pöppel et al.

ICLR 2025arXiv:2406.04303
87
citations
#518

Amortizing intractable inference in large language models

Edward Hu, Moksh Jain, Eric Elmoznino et al.

ICLR 2024arXiv:2310.04363
86
citations
#519

MM-EMBED: UNIVERSAL MULTIMODAL RETRIEVAL WITH MULTIMODAL LLMS

Sheng-Chieh Lin, Chankyu Lee, Mohammad Shoeybi et al.

ICLR 2025arXiv:2411.02571
86
citations
#520

Safety Layers in Aligned Large Language Models: The Key to LLM Security

Shen Li, Liuyi Yao, Lan Zhang et al.

ICLR 2025arXiv:2408.17003
86
citations
#521

Robust Watermarking Using Generative Priors Against Image Editing: From Benchmarking to Advances

Shilin Lu, Zihan Zhou, Jiayou Lu et al.

ICLR 2025arXiv:2410.18775
86
citations
#522

MotionClone: Training-Free Motion Cloning for Controllable Video Generation

Pengyang Ling, Jiazi Bu, Pan Zhang et al.

ICLR 2025oralarXiv:2406.05338
86
citations
#523

Neural Common Neighbor with Completion for Link Prediction

Xiyuan Wang, Haotong Yang, Muhan Zhang

ICLR 2024arXiv:2302.00890
86
citations
#524

AHA: A Vision-Language-Model for Detecting and Reasoning Over Failures in Robotic Manipulation

Jiafei Duan, Wilbert Pumacay, Nishanth Kumar et al.

ICLR 2025arXiv:2410.00371
85
citations
#525

Programming Refusal with Conditional Activation Steering

Bruce W. Lee, Inkit Padhi, Karthikeyan Natesan Ramamurthy et al.

ICLR 2025arXiv:2409.05907
85
citations
#526

Large-scale Training of Foundation Models for Wearable Biosignals

Salar Abbaspourazad, Oussama Elachqar, Andrew Miller et al.

ICLR 2024arXiv:2312.05409
85
citations
#527

Do I Know This Entity? Knowledge Awareness and Hallucinations in Language Models

Javier Ferrando, Oscar Obeso, Senthooran Rajamanoharan et al.

ICLR 2025arXiv:2411.14257
85
citations
#528

Improved Techniques for Optimization-Based Jailbreaking on Large Language Models

Xiaojun Jia, Tianyu Pang, Chao Du et al.

ICLR 2025arXiv:2405.21018
85
citations
#529

LQ-LoRA: Low-rank plus Quantized Matrix Decomposition for Efficient Language Model Finetuning

Han Guo, Philip Greengard, Eric Xing et al.

ICLR 2024arXiv:2311.12023
85
citations
#530

Human Feedback is not Gold Standard

Tom Hosking, Phil Blunsom, Max Bartolo

ICLR 2024arXiv:2309.16349
84
citations
#531

BadChain: Backdoor Chain-of-Thought Prompting for Large Language Models

Zhen Xiang, Fengqing Jiang, Zidi Xiong et al.

ICLR 2024arXiv:2401.12242
84
citations
#532

Controlling Vision-Language Models for Multi-Task Image Restoration

Ziwei Luo, Fredrik K. Gustafsson, Zheng Zhao et al.

ICLR 2024arXiv:2310.01018
84
citations
#533

SuRe: Summarizing Retrievals using Answer Candidates for Open-domain QA of LLMs

Jaehyung Kim, Jaehyun Nam, Sangwoo Mo et al.

ICLR 2024arXiv:2404.13081
84
citations
#534

AntGPT: Can Large Language Models Help Long-term Action Anticipation from Videos?

Qi Zhao, Shijie Wang, Ce Zhang et al.

ICLR 2024oralarXiv:2307.16368
84
citations
#535

Beyond Autoregression: Discrete Diffusion for Complex Reasoning and Planning

Jiacheng Ye, Jiahui Gao, Shansan Gong et al.

ICLR 2025arXiv:2410.14157
84
citations
#536

Training-free Camera Control for Video Generation

Chen Hou, Zhibo Chen

ICLR 2025arXiv:2406.10126
84
citations
#537

MMed-RAG: Versatile Multimodal RAG System for Medical Vision Language Models

Peng Xia, Kangyu Zhu, Haoran Li et al.

ICLR 2025arXiv:2410.13085
83
citations
#538

Skeleton-of-Thought: Prompting LLMs for Efficient Parallel Generation

Xuefei Ning, Zinan Lin, Zixuan Zhou et al.

ICLR 2024arXiv:2307.15337
83
citations
#539

CFG++: Manifold-constrained Classifier Free Guidance for Diffusion Models

Hyungjin Chung, Jeongsol Kim, Geon Yeong Park et al.

ICLR 2025arXiv:2406.08070
83
citations
#540

CARD: Channel Aligned Robust Blend Transformer for Time Series Forecasting

xue wang, Tian Zhou, Qingsong Wen et al.

ICLR 2024oralarXiv:2305.12095
83
citations
#541

ClimODE: Climate and Weather Forecasting with Physics-informed Neural ODEs

Yogesh Verma, Markus Heinonen, Vikas Garg

ICLR 2024oralarXiv:2404.10024
82
citations
#542

Soft Merging of Experts with Adaptive Routing

Haokun Liu, Muqeeth Mohammed, Colin Raffel

ICLR 2025arXiv:2306.03745
82
citations
#543

Diffusion-Based Planning for Autonomous Driving with Flexible Guidance

Yinan Zheng, Ruiming Liang, Kexin ZHENG et al.

ICLR 2025arXiv:2501.15564
82
citations
#544

PB-LLM: Partially Binarized Large Language Models

Zhihang Yuan, Yuzhang Shang, Zhen Dong

ICLR 2024arXiv:2310.00034
82
citations
#545

Towards Foundation Models for Knowledge Graph Reasoning

Mikhail Galkin, Xinyu Yuan, Hesham Mostafa et al.

ICLR 2024arXiv:2310.04562
82
citations
#546

In-Context Pretraining: Language Modeling Beyond Document Boundaries

Weijia Shi, Sewon Min, Maria Lomeli et al.

ICLR 2024spotlightarXiv:2310.10638
81
citations
#547

Real-Time Video Generation with Pyramid Attention Broadcast

Xuanlei Zhao, Xiaolong Jin, Kai Wang et al.

ICLR 2025arXiv:2408.12588
81
citations
#548

Convolution Meets LoRA: Parameter Efficient Finetuning for Segment Anything Model

Zihan Zhong, Zhiqiang Tang, Tong He et al.

ICLR 2024arXiv:2401.17868
81
citations
#549

Language Model Self-improvement by Reinforcement Learning Contemplation

Jing-Cheng Pang, Pengyuan Wang, Kaiyuan Li et al.

ICLR 2024arXiv:2305.14483
81
citations
#550

Dissecting Adversarial Robustness of Multimodal LM Agents

Chen Wu, Rishi Shah, Jing Yu Koh et al.

ICLR 2025arXiv:2406.12814
81
citations
#551

A Benchmark for Learning to Translate a New Language from One Grammar Book

Garrett Tanzer, Mirac Suzgun, Eline Visser et al.

ICLR 2024spotlightarXiv:2309.16575
81
citations
#552

D-FINE: Redefine Regression Task of DETRs as Fine-grained Distribution Refinement

Yansong Peng, Hebei Li, Peixi Wu et al.

ICLR 2025arXiv:2410.13842
81
citations
#553

DyVal: Dynamic Evaluation of Large Language Models for Reasoning Tasks

Kaijie Zhu, Jiaao Chen, Jindong Wang et al.

ICLR 2024spotlightarXiv:2309.17167
81
citations
#554

Hallo2: Long-Duration and High-Resolution Audio-Driven Portrait Image Animation

Jiahao Cui, Hui Li, Yao Yao et al.

ICLR 2025oralarXiv:2410.07718
80
citations
#555

InfoBatch: Lossless Training Speed Up by Unbiased Dynamic Data Pruning

Ziheng Qin, Kai Wang, Zangwei Zheng et al.

ICLR 2024arXiv:2303.04947
80
citations
#556

Improving Instruction-Following in Language Models through Activation Steering

Alessandro Stolfo, Vidhisha Balachandran, Safoora Yousefi et al.

ICLR 2025arXiv:2410.12877
80
citations
#557

MMTEB: Massive Multilingual Text Embedding Benchmark

Kenneth Enevoldsen, Isaac Chung, Imene Kerboua et al.

ICLR 2025arXiv:2502.13595
80
citations
#558

DeepZero: Scaling Up Zeroth-Order Optimization for Deep Model Training

AOCHUAN CHEN, Yimeng Zhang, Jinghan Jia et al.

ICLR 2024arXiv:2310.02025
80
citations
#559

Generalization v.s. Memorization: Tracing Language Models’ Capabilities Back to Pretraining Data

Xinyi Wang, Antonis Antoniades, Yanai Elazar et al.

ICLR 2025arXiv:2407.14985
80
citations
#560

Exploratory Preference Optimization: Harnessing Implicit Q*-Approximation for Sample-Efficient RLHF

Tengyang Xie, Dylan Foster, Akshay Krishnamurthy et al.

ICLR 2025arXiv:2405.21046
79
citations
#561

Linear attention is (maybe) all you need (to understand Transformer optimization)

Kwangjun Ahn, Xiang Cheng, Minhak Song et al.

ICLR 2024arXiv:2310.01082
79
citations
#562

Unified Language-Vision Pretraining in LLM with Dynamic Discrete Visual Tokenization

Yang Jin, Kun Xu, Kun Xu et al.

ICLR 2024arXiv:2309.04669
79
citations
#563

On the Stability of Iterative Retraining of Generative Models on their own Data

Quentin Bertrand, Joey Bose, Alexandre Duplessis et al.

ICLR 2024spotlightarXiv:2310.00429
79
citations
#564

How to Catch an AI Liar: Lie Detection in Black-Box LLMs by Asking Unrelated Questions

Lorenzo Pacchiardi, Alex Chan, Sören Mindermann et al.

ICLR 2024arXiv:2309.15840
79
citations
#565

MAVIS: Mathematical Visual Instruction Tuning with an Automatic Data Engine

Renrui Zhang, Xinyu Wei, Dongzhi Jiang et al.

ICLR 2025arXiv:2407.08739
79
citations
#566

Towards 3D Molecule-Text Interpretation in Language Models

Sihang Li, Zhiyuan Liu, Yanchen Luo et al.

ICLR 2024arXiv:2401.13923
79
citations
#567

Curiosity-driven Red-teaming for Large Language Models

Zhang-Wei Hong, Idan Shenfeld, Johnson (Tsun-Hsuan) Wang et al.

ICLR 2024arXiv:2402.19464
79
citations
#568

Eliciting Human Preferences with Language Models

Belinda Li, Alex Tamkin, Noah Goodman et al.

ICLR 2025oralarXiv:2310.11589
79
citations
#569

Language models scale reliably with over-training and on downstream tasks

Samir Yitzhak Gadre, Georgios Smyrnis, Vaishaal Shankar et al.

ICLR 2025arXiv:2403.08540
79
citations
#570

Reasoning with Latent Thoughts: On the Power of Looped Transformers

Nikunj Saunshi, Nishanth Dikkala, Zhiyuan Li et al.

ICLR 2025arXiv:2502.17416
79
citations
#571

Learning to Act without Actions

Dominik Schmidt, Minqi Jiang

ICLR 2024oralarXiv:2312.10812
78
citations
#572

Language Models Learn to Mislead Humans via RLHF

Jiaxin Wen, Ruiqi Zhong, Akbir Khan et al.

ICLR 2025arXiv:2409.12822
78
citations
#573

Speculative RAG: Enhancing Retrieval Augmented Generation through Drafting

Zilong (Ryan) Wang, Zifeng Wang, Long Le et al.

ICLR 2025arXiv:2407.08223
78
citations
#574

Remote Sensing Vision-Language Foundation Models without Annotations via Ground Remote Alignment

Utkarsh Kumar Mall, Cheng Perng Phoo, Meilin Liu et al.

ICLR 2024arXiv:2312.06960
78
citations
#575

RoboCat: A Self-Improving Generalist Agent for Robotic Manipulation

Sergio Gómez Colmenarejo, Jost Springenberg, Jose Enrique Chen et al.

ICLR 2025
78
citations
#576

Multiscale Positive-Unlabeled Detection of AI-Generated Texts

Yuchuan Tian, Hanting Chen, Xutao Wang et al.

ICLR 2024spotlightarXiv:2305.18149
78
citations
#577

DreamTime: An Improved Optimization Strategy for Diffusion-Guided 3D Generation

Yukun Huang, Jianan Wang, Yukai Shi et al.

ICLR 2024arXiv:2306.12422
78
citations
#578

Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages

Jinyi Hu, Yuan Yao, Chongyi Wang et al.

ICLR 2024spotlightarXiv:2308.12038
77
citations
#579

Elucidating the Exposure Bias in Diffusion Models

Mang Ning, Mingxiao Li, Jianlin Su et al.

ICLR 2024arXiv:2308.15321
77
citations
#580

Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy

Pingzhi Li, Zhenyu Zhang, Prateek Yadav et al.

ICLR 2024spotlightarXiv:2310.01334
77
citations
#581

Robustness of AI-Image Detectors: Fundamental Limits and Practical Attacks

Mehrdad Saberi, Vinu Sankar Sadasivan, Keivan Rezaei et al.

ICLR 2024arXiv:2310.00076
77
citations
#582

Single Motion Diffusion

Sigal Raab, Inbal Leibovitch, Guy Tevet et al.

ICLR 2024oralarXiv:2302.05905
77
citations
#583

FakeShield: Explainable Image Forgery Detection and Localization via Multi-modal Large Language Models

Zhipei Xu, Xuanyu Zhang, Runyi Li et al.

ICLR 2025arXiv:2410.02761
77
citations
#584

How Do Transformers Learn In-Context Beyond Simple Functions? A Case Study on Learning with Representations

Tianyu Guo, Wei Hu, Song Mei et al.

ICLR 2024arXiv:2310.10616
77
citations
#585

Model Merging by Uncertainty-Based Gradient Matching

Nico Daheim, Thomas Möllenhoff, Edoardo M. Ponti et al.

ICLR 2024arXiv:2310.12808
77
citations
#586

LLM-grounded Video Diffusion Models

Long Lian, Baifeng Shi, Adam Yala et al.

ICLR 2024oralarXiv:2309.17444
77
citations
#587

Enhancing End-to-End Autonomous Driving with Latent World Model

Yingyan Li, Lue Fan, Jiawei He et al.

ICLR 2025arXiv:2406.08481
77
citations
#588

Test of Time: A Benchmark for Evaluating LLMs on Temporal Reasoning

Bahare Fatemi, Seyed Mehran Kazemi, Anton Tsitsulin et al.

ICLR 2025oralarXiv:2406.09170
76
citations
#589

InstructRAG: Instructing Retrieval-Augmented Generation via Self-Synthesized Rationales

Zhepei Wei, Wei-Lin Chen, Yu Meng

ICLR 2025arXiv:2406.13629
76
citations
#590

Simple Guidance Mechanisms for Discrete Diffusion Models

Yair Schiff, Subham Sahoo, Hao Phung et al.

ICLR 2025arXiv:2412.10193
76
citations
#591

Round and Round We Go! What makes Rotary Positional Encodings useful?

Federico Barbero, Alex Vitvitskyi, Christos Perivolaropoulos et al.

ICLR 2025arXiv:2410.06205
76
citations
#592

Fine-tuning can cripple your foundation model; preserving features may be the solution

Philip Torr, Puneet Dokania, Jishnu Mukhoti et al.

ICLR 2025arXiv:2308.13320
76
citations
#593

Plan-Seq-Learn: Language Model Guided RL for Solving Long Horizon Robotics Tasks

Murtaza Dalal, Tarun Chiruvolu, Devendra Chaplot et al.

ICLR 2024arXiv:2405.01534
76
citations
#594

MaskBit: Embedding-free Image Generation via Bit Tokens

Mark Weber, Lijun Yu, Qihang Yu et al.

ICLR 2025arXiv:2409.16211
75
citations
#595

Confronting Reward Model Overoptimization with Constrained RLHF

Ted Moskovitz, Aaditya Singh, DJ Strouse et al.

ICLR 2024spotlightarXiv:2310.04373
75
citations
#596

HyperHuman: Hyper-Realistic Human Generation with Latent Structural Diffusion

Xian Liu, Jian Ren, Aliaksandr Siarohin et al.

ICLR 2024arXiv:2310.08579
75
citations
#597

METRA: Scalable Unsupervised RL with Metric-Aware Abstraction

Seohong Park, Oleh Rybkin, Sergey Levine

ICLR 2024oralarXiv:2310.08887
75
citations
#598

HAMSTER: Hierarchical Action Models for Open-World Robot Manipulation

Yi Li, Yuquan Deng, Jesse Zhang et al.

ICLR 2025arXiv:2502.05485
75
citations
#599

LLM-CXR: Instruction-Finetuned LLM for CXR Image Understanding and Generation

Suhyeon Lee, Won Jun Kim, Jinho Chang et al.

ICLR 2024arXiv:2305.11490
75
citations
#600

Understanding In-Context Learning in Transformers and LLMs by Learning to Learn Discrete Functions

Satwik Bhattamishra, Arkil Patel, Phil Blunsom et al.

ICLR 2024arXiv:2310.03016
75
citations