ICLR Papers

6,124 papers found • Page 76 of 123

Weakly Supervised Video Scene Graph Generation via Natural Language Supervision

Kibum Kim, Kanghoon Yoon, Yeonjun In et al.

ICLR 2025oralarXiv:2502.15370
2
citations

Weak to Strong Generalization for Large Language Models with Multi-capabilities

Yucheng Zhou, Jianbing Shen, Yu Cheng

ICLR 2025poster
70
citations

Weak-to-Strong Generalization Through the Data-Centric Lens

Changho Shin, John Cooper, Frederic Sala

ICLR 2025posterarXiv:2412.03881
14
citations

Weak-to-Strong Preference Optimization: Stealing Reward from Weak Aligned Model

Wenhong Zhu, Zhiwei He, Xiaofeng Wang et al.

ICLR 2025posterarXiv:2410.18640
14
citations

WeatherGFM: Learning a Weather Generalist Foundation Model via In-context Learning

Xiangyu Zhao, Zhiwang Zhou, Wenlong Zhang et al.

ICLR 2025oralarXiv:2411.05420
9
citations

Web Agents with World Models: Learning and Leveraging Environment Dynamics in Web Navigation

Hyungjoo Chae, Namyoung Kim, Kai Ong et al.

ICLR 2025posterarXiv:2410.13232
59
citations

WebRL: Training LLM Web Agents via Self-Evolving Online Curriculum Reinforcement Learning

Zehan Qi, Xiao Liu, Iat Long Iong et al.

ICLR 2025posterarXiv:2411.02337
117
citations

Weighted Multi-Prompt Learning with Description-free Large Language Model Distillation

Sua Lee, Kyubum Shin, Jung Ho Park

ICLR 2025posterarXiv:2507.07147
1
citations

Weighted Point Set Embedding for Multimodal Contrastive Learning Toward Optimal Similarity Metric

Toshimitsu Uesaka, Taiji Suzuki, Yuhta Takida et al.

ICLR 2025posterarXiv:2404.19228
3
citations

Weighted-Reward Preference Optimization for Implicit Model Fusion

Ziyi Yang, Fanqi Wan, Longguang Zhong et al.

ICLR 2025posterarXiv:2412.03187
14
citations

What Are Good Positional Encodings for Directed Graphs?

Yinan Huang, Haoyu Wang, Pan Li

ICLR 2025posterarXiv:2407.20912

What Does It Mean to Be a Transformer? Insights from a Theoretical Hessian Analysis

Weronika Ormaniec, Felix Dangel, Sidak Pal Singh

ICLR 2025posterarXiv:2410.10986
10
citations

What Do You See in Common? Learning Hierarchical Prototypes over Tree-of-Life to Discover Evolutionary Traits

Harish Babu Manogaran, M. Maruf, Arka Daw et al.

ICLR 2025posterarXiv:2409.02335
1
citations

What Has Been Overlooked in Contrastive Source-Free Domain Adaptation: Leveraging Source-Informed Latent Augmentation within Neighborhood Context

JING WANG, Wonho Bae, Jiahong Chen et al.

ICLR 2025posterarXiv:2412.14301
7
citations

What is Wrong with Perplexity for Long-context Language Modeling?

Lizhe Fang, Yifei Wang, Zhaoyang Liu et al.

ICLR 2025posterarXiv:2410.23771

What Makes a Good Diffusion Planner for Decision Making?

Haofei Lu, Dongqi Han, Yifei Shen et al.

ICLR 2025posterarXiv:2503.00535
26
citations

What Makes a Maze Look Like a Maze?

Joy Hsu, Jiayuan Mao, Joshua B Tenenbaum et al.

ICLR 2025posterarXiv:2409.08202
13
citations

What Makes Large Language Models Reason in (Multi-Turn) Code Generation?

Kunhao Zheng, Juliette Decugis, Jonas Gehring et al.

ICLR 2025posterarXiv:2410.08105
32
citations

What Matters in Learning from Large-Scale Datasets for Robot Manipulation

Vaibhav Saxena, Matthew Bronars, Nadun Ranawaka Arachchige et al.

ICLR 2025posterarXiv:2506.13536
16
citations

What Matters When Repurposing Diffusion Models for General Dense Perception Tasks?

Guangkai Xu, yongtao ge, Mingyu Liu et al.

ICLR 2025posterarXiv:2403.06090
56
citations

What Secrets Do Your Manifolds Hold? Understanding the Local Geometry of Generative Models

Ahmed Imtiaz Humayun, Ibtihel Amara, Cristina Nader Vasconcelos et al.

ICLR 2025posterarXiv:2408.08307

What should a neuron aim for? Designing local objective functions based on information theory

Andreas C. Schneider, Valentin Neuhaus, David Ehrlich et al.

ICLR 2025posterarXiv:2412.02482
5
citations

What's New in My Data? Novelty Exploration via Contrastive Generation

Masaru Isonuma, Ivan Titov

ICLR 2025posterarXiv:2410.14765

What's the Move? Hybrid Imitation Learning via Salient Points

Priya Sundaresan, Hengyuan Hu, Quan Vuong et al.

ICLR 2025posterarXiv:2412.05426
11
citations

What to align in multimodal contrastive learning?

Benoit Dufumier, Javiera Castillo Navarro, Devis Tuia et al.

ICLR 2025posterarXiv:2409.07402
33
citations

When Attention Sink Emerges in Language Models: An Empirical View

Xiangming Gu, Tianyu Pang, Chao Du et al.

ICLR 2025posterarXiv:2410.10781
90
citations

When does compositional structure yield compositional generalization? A kernel theory.

Samuel Lippl, Kimberly Stachenfeld

ICLR 2025posterarXiv:2405.16391

When do GFlowNets learn the right distribution?

Tiago Silva, Rodrigo Alves, Eliezer de Souza da Silva et al.

ICLR 2025poster

When GNNs meet symmetry in ILPs: an orbit-based feature augmentation approach

Qian Chen, Lei Li, Qian Li et al.

ICLR 2025posterarXiv:2501.14211
1
citations

When Graph Neural Networks Meet Dynamic Mode Decomposition

Dai Shi, Lequan Lin, Andi Han et al.

ICLR 2025oralarXiv:2410.05593

When is Task Vector Provably Effective for Model Editing? A Generalization Analysis of Nonlinear Transformers

Hongkang Li, Yihua Zhang, shuai ZHANG et al.

ICLR 2025posterarXiv:2504.10957

When LLMs Play the Telephone Game: Cultural Attractors as Conceptual Tools to Evaluate LLMs in Multi-turn Settings

Jérémy Perez, Grgur Kovac, Corentin Léger et al.

ICLR 2025posterarXiv:2407.04503
5
citations

When narrower is better: the narrow width limit of Bayesian parallel branching neural networks

Zechen Zhang, Haim Sompolinsky

ICLR 2025posterarXiv:2407.18807
1
citations

When Prompt Engineering Meets Software Engineering: CNL-P as Natural and Robust "APIs'' for Human-AI Interaction

Zhenchang Xing, Yang Liu, Zhuo Cheng et al.

ICLR 2025poster

When Selection Meets Intervention: Additional Complexities in Causal Discovery

Haoyue Dai, Ignavier Ng, Jianle Sun et al.

ICLR 2025posterarXiv:2503.07302
5
citations

Where Am I and What Will I See: An Auto-Regressive Model for Spatial Localization and View Prediction

Junyi Chen, Di Huang, Weicai Ye et al.

ICLR 2025posterarXiv:2410.18962
4
citations

Which Tasks Should Be Compressed Together? A Causal Discovery Approach for Efficient Multi-Task Representation Compression

Sha Guo, Jing Chen, Zixuan Hu et al.

ICLR 2025poster
1
citations

Why Does the Effective Context Length of LLMs Fall Short?

Chenxin An, Jun Zhang, Ming Zhong et al.

ICLR 2025posterarXiv:2410.18745

Why In-Context Learning Models are Good Few-Shot Learners?

Shiguang Wu, Yaqing Wang, Quanming Yao

ICLR 2025poster

Why RoPE Struggles to Maintain Long-Term Decay in Long Sequences?

Wei Shen, Chao Yin, Yuliang Liu et al.

ICLR 2025poster

Wicked Oddities: Selectively Poisoning for Effective Clean-Label Backdoor Attacks

Hung Quang Nguyen, Hieu Nguyen, Anh Ta et al.

ICLR 2025posterarXiv:2407.10825

Wide Neural Networks Trained with Weight Decay Provably Exhibit Neural Collapse

Arthur Jacot, Peter Súkeník, Zihan Wang et al.

ICLR 2025posterarXiv:2410.04887
9
citations

WildBench: Benchmarking LLMs with Challenging Tasks from Real Users in the Wild

Bill Yuchen Lin, Yuntian Deng, Khyathi Chandu et al.

ICLR 2025posterarXiv:2406.04770
142
citations

WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct

Haipeng Luo, Qingfeng Sun, Can Xu et al.

ICLR 2025posterarXiv:2308.09583
644
citations

Words in Motion: Extracting Interpretable Control Vectors for Motion Transformers

Omer Sahin Tas, Royden Wagner

ICLR 2025posterarXiv:2406.11624
4
citations

WorkflowLLM: Enhancing Workflow Orchestration Capability of Large Language Models

Shengda Fan, Xin Cong, Yuepeng Fu et al.

ICLR 2025posterarXiv:2411.05451
14
citations

World Model on Million-Length Video And Language With Blockwise RingAttention

Hao Liu, Wilson Yan, Matei Zaharia et al.

ICLR 2025oralarXiv:2402.08268
144
citations

W-PCA Based Gradient-Free Proxy for Efficient Search of Lightweight Language Models

Shang Wang

ICLR 2025posterarXiv:2504.15983

XAIguiFormer: explainable artificial intelligence guided transformer for brain disorder identification

Hanning Guo, Farah Abdellatif, Yu Fu et al.

ICLR 2025poster

X-ALMA: Plug & Play Modules and Adaptive Rejection for Quality Translation at Scale

Haoran Xu, Kenton Murray, Philipp Koehn et al.

ICLR 2025posterarXiv:2410.03115