ICLR Poster "training efficiency" Papers
4 papers found
Cut Your Losses in Large-Vocabulary Language Models
Erik Wijmans, Brody Huval, Alexander Hertzberg et al.
ICLR 2025posterarXiv:2411.09009
19
citations
Fewer May Be Better: Enhancing Offline Reinforcement Learning with Reduced Dataset
Yiqin Yang, Quanwei Wang, Chenghao Li et al.
ICLR 2025posterarXiv:2502.18955
Linear Combination of Saved Checkpoints Makes Consistency and Diffusion Models Better
Enshu Liu, Junyi Zhu, Zinan Lin et al.
ICLR 2025posterarXiv:2404.02241
6
citations
Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think
Sihyun Yu, Sangkyung Kwak, Huiwon Jang et al.
ICLR 2025posterarXiv:2410.06940
308
citations