ICLR Poster Papers

5,330 papers found • Page 18 of 107

DistRL: An Asynchronous Distributed Reinforcement Learning Framework for On-Device Control Agent

Taiyi Wang, Zhihao Wu, Jianheng Liu et al.

ICLR 2025posterarXiv:2410.14803

DiTTo-TTS: Diffusion Transformers for Scalable Text-to-Speech without Domain-Specific Factors

Keon Lee, Dong Won Kim, Jaehyeon Kim et al.

ICLR 2025posterarXiv:2406.11427
28
citations

Divergence-enhanced Knowledge-guided Context Optimization for Visual-Language Prompt Tuning

Yilun Li, Miaomiao Cheng, Xu Han et al.

ICLR 2025poster
6
citations

Divergence of Neural Tangent Kernel in Classification Problems

Zixiong Yu, Songtao Tian, Guhan Chen

ICLR 2025poster

Divergence-Regularized Discounted Aggregation: Equilibrium Finding in Multiplayer Partially Observable Stochastic Games

Runyu Lu, Yuanheng Zhu, Dongbin Zhao

ICLR 2025poster
3
citations

Diverse Preference Learning for Capabilities and Alignment

Stewart Slocum, Asher Parker-Sartori, Dylan Hadfield-Menell

ICLR 2025posterarXiv:2511.08594
21
citations

Diversity Empowers Intelligence: Integrating Expertise of Software Engineering Agents

Kexun Zhang, Weiran Yao, Zuxin Liu et al.

ICLR 2025posterarXiv:2408.07060
39
citations

Diversity-Rewarded CFG Distillation

Geoffrey Cideron, Andrea Agostinelli, Johan Ferret et al.

ICLR 2025posterarXiv:2410.06084

Divide and Translate: Compositional First-Order Logic Translation and Verification for Complex Logical Reasoning

Hyun Ryu, Gyeongman Kim, Hyemin S. Lee et al.

ICLR 2025posterarXiv:2410.08047
25
citations

DLEFT-MKC: Dynamic Late Fusion Multiple Kernel Clustering with Robust Tensor Learning via Min-Max Optimization

Yi Zhang, Siwei Wang, Jiyuan Liu et al.

ICLR 2025poster
1
citations

Do as I do (Safely): Mitigating Task-Specific Fine-tuning Risks in Large Language Models

Francisco Eiras, Aleksandar Petrov, Philip Torr et al.

ICLR 2025posterarXiv:2406.10288

Do as We Do, Not as You Think: the Conformity of Large Language Models

Zhiyuan Weng, Guikun Chen, Wenguan Wang

ICLR 2025posterarXiv:2501.13381
18
citations

Dobi-SVD: Differentiable SVD for LLM Compression and Some New Perspectives

Qinsi Wang, Jinghan Ke, Masayoshi Tomizuka et al.

ICLR 2025posterarXiv:2502.02723

DocMIA: Document-Level Membership Inference Attacks against DocVQA Models

Khanh Nguyen, Raouf Kerkouche, Mario Fritz et al.

ICLR 2025posterarXiv:2502.03692

Do Contemporary Causal Inference Models Capture Real-World Heterogeneity? Findings from a Large-Scale Benchmark

Haining Yu, Yizhou Sun

ICLR 2025posterarXiv:2410.07021

DOCS: Quantifying Weight Similarity for Deeper Insights into Large Language Models

Zeping Min, Xinshang Wang

ICLR 2025posterarXiv:2501.16650

Do Deep Neural Network Solutions Form a Star Domain?

Ankit Sonthalia, Alexander Rubinstein, Ehsan Abbasnejad et al.

ICLR 2025posterarXiv:2403.07968
4
citations

Does Editing Provide Evidence for Localization?

Zihao Wang, Victor Veitch

ICLR 2025posterarXiv:2502.11447
9
citations

Does Refusal Training in LLMs Generalize to the Past Tense?

Maksym Andriushchenko, Nicolas Flammarion

ICLR 2025posterarXiv:2407.11969
66
citations

Does Safety Training of LLMs Generalize to Semantically Related Natural Prompts?

Sravanti Addepalli, Yerram Varun, Arun Suggala et al.

ICLR 2025posterarXiv:2412.03235
7
citations

Does SGD really happen in tiny subspaces?

Minhak Song, Kwangjun Ahn, Chulhee Yun

ICLR 2025posterarXiv:2405.16002
16
citations

Does Spatial Cognition Emerge in Frontier Models?

Santhosh Kumar Ramakrishnan, Erik Wijmans, Philipp Krähenbühl et al.

ICLR 2025posterarXiv:2410.06468
51
citations

Does Training with Synthetic Data Truly Protect Privacy?

Yunpeng Zhao, Jie Zhang

ICLR 2025posterarXiv:2502.12976

DoF: A Diffusion Factorization Framework for Offline Multi-Agent Reinforcement Learning

Chao Li, Ziwei Deng, Chenxing Lin et al.

ICLR 2025poster
7
citations

Do I Know This Entity? Knowledge Awareness and Hallucinations in Language Models

Javier Ferrando, Oscar Obeso, Senthooran Rajamanoharan et al.

ICLR 2025posterarXiv:2411.14257
77
citations

Do Large Language Models Truly Understand Geometric Structures?

Xiaofeng Wang, Yiming Wang, Wenhong Zhu et al.

ICLR 2025posterarXiv:2501.13773
9
citations

Do LLM Agents Have Regret? A Case Study in Online Learning and Games

Chanwoo Park, Xiangyu Liu, Asuman Ozdaglar et al.

ICLR 2025posterarXiv:2403.16843

Do LLMs estimate uncertainty well in instruction-following?

Juyeon Heo, Miao Xiong, Christina Heinze-Deml et al.

ICLR 2025posterarXiv:2410.14582
13
citations

Do LLMs have Consistent Values?

Naama Rozen, Liat Bezalel, Gal Elidan et al.

ICLR 2025posterarXiv:2407.12878

Do LLMs ``know'' internally when they follow instructions?

Juyeon Heo, Christina Heinze-Deml, Oussama Elachqar et al.

ICLR 2025posterarXiv:2410.14516
22
citations

Do LLMs Recognize Your Preferences? Evaluating Personalized Preference Following in LLMs

Siyan Zhao, Mingyi Hong, Yang Liu et al.

ICLR 2025posterarXiv:2502.09597

Domain Guidance: A Simple Transfer Approach for a Pre-trained Diffusion Model

Jincheng Zhong, XiangCheng Zhang, Jianmin Wang et al.

ICLR 2025posterarXiv:2504.01521
4
citations

Do Mice Grok? Glimpses of Hidden Progress in Sensory Cortex

Tanishq Kumar, Blake Bordelon, Cengiz Pehlevan et al.

ICLR 2025poster
1
citations

Do not write that jailbreak paper

Javier Rando

ICLR 2025poster

Don't flatten, tokenize! Unlocking the key to SoftMoE's efficacy in deep RL

Ghada Sokar, Johan S Obando Ceron, Aaron Courville et al.

ICLR 2025posterarXiv:2410.01930

DON’T STOP ME NOW: EMBEDDING BASED SCHEDULING FOR LLMS

Rana Shahout, Eran Malach, Chunwei Liu et al.

ICLR 2025poster
15
citations

Don't Take Things Out of Context: Attention Intervention for Enhancing Chain-of-Thought Reasoning in Large Language Models

Shaotian Yan, Chen Shen, Wenxiao Wang et al.

ICLR 2025posterarXiv:2503.11154

DOPL: Direct Online Preference Learning for Restless Bandits with Preference Feedback

GUOJUN XIONG, Ujwal Dinesha, Debajoy Mukherjee et al.

ICLR 2025posterarXiv:2410.05527

Do Stochastic, Feel Noiseless: Stable Stochastic Optimization via a Double Momentum Mechanism

Tehila Dahan, Kfir Y Levy

ICLR 2025poster

DOTS: Learning to Reason Dynamically in LLMs via Optimal Reasoning Trajectories Search

Murong Yue, Wenlin Yao, Haitao Mi et al.

ICLR 2025posterarXiv:2410.03864

Doubly Optimal Policy Evaluation for Reinforcement Learning

Shuze Liu, Claire Chen, Shangtong Zhang

ICLR 2025posterarXiv:2410.02226
3
citations

Doubly robust identification of treatment effects from multiple environments

Piersilvio De Bartolomeis, Julia Kostin, Javier Abad et al.

ICLR 2025posterarXiv:2503.14459

Do Vision & Language Decoders use Images and Text equally? How Self-consistent are their Explanations?

Letitia Parcalabescu, Anette Frank

ICLR 2025posterarXiv:2404.18624
20
citations

Do Vision-Language Models Represent Space and How? Evaluating Spatial Frame of Reference under Ambiguities

Zheyuan Zhang, Fengyuan Hu, Jayjun Lee et al.

ICLR 2025posterarXiv:2410.17385
40
citations

Do vision models perceive objects like toddlers ?

Arthur Aubret, Jochen Triesch

ICLR 2025poster

Do WGANs succeed because they minimize the Wasserstein Distance? Lessons from Discrete Generators

Ariel Elnekave, Yair Weiss

ICLR 2025poster

Do You Keep an Eye on What I Ask? Mitigating Multimodal Hallucination via Attention-Guided Ensemble Decoding

Yeongjae Cho, Keonwoo Kim, Taebaek Hwang et al.

ICLR 2025posterarXiv:2505.17529

DPaI: Differentiable Pruning at Initialization with Node-Path Balance Principle

Lichuan Xiang, Quan Nguyen-Tri, Lan-Cuong Nguyen et al.

ICLR 2025poster

DPLM-2: A Multimodal Diffusion Protein Language Model

Xinyou Wang, Zaixiang Zheng, Fei YE et al.

ICLR 2025posterarXiv:2410.13782

Drama: Mamba-Enabled Model-Based Reinforcement Learning Is Sample and Parameter Efficient

Wenlong Wang, Ivana Dusparic, Yucheng Shi et al.

ICLR 2025posterarXiv:2410.08893
3
citations