"training efficiency" Papers
12 papers found
Cut Your Losses in Large-Vocabulary Language Models
Erik Wijmans, Brody Huval, Alexander Hertzberg et al.
ICLR 2025posterarXiv:2411.09009
19
citations
Efficient Representativeness-Aware Coreset Selection
Zihao Cheng, Binrui Wu, Zhiwei Li et al.
NeurIPS 2025poster
Faster and Better 3D Splatting via Group Training
Chengbo Wang, Guozheng Ma, Yizhen Lao et al.
ICCV 2025posterarXiv:2412.07608
3
citations
Fewer May Be Better: Enhancing Offline Reinforcement Learning with Reduced Dataset
Yiqin Yang, Quanwei Wang, Chenghao Li et al.
ICLR 2025posterarXiv:2502.18955
Linear Combination of Saved Checkpoints Makes Consistency and Diffusion Models Better
Enshu Liu, Junyi Zhu, Zinan Lin et al.
ICLR 2025posterarXiv:2404.02241
6
citations
Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think
Sihyun Yu, Sangkyung Kwak, Huiwon Jang et al.
ICLR 2025posterarXiv:2410.06940
308
citations
Representation Entanglement for Generation: Training Diffusion Transformers Is Much Easier Than You Think
Ge Wu, Shen Zhang, Ruijing Shi et al.
NeurIPS 2025oralarXiv:2507.01467
27
citations
BWS: Best Window Selection Based on Sample Scores for Data Pruning across Broad Ranges
Hoyong Choi, Nohyun Ki, Hye Won Chung
ICML 2024poster
Diversified Batch Selection for Training Acceleration
Feng Hong, Yueming LYU, Jiangchao Yao et al.
ICML 2024poster
EE-LLM: Large-Scale Training and Inference of Early-Exit Large Language Models with 3D Parallelism
Yanxi Chen, Xuchen Pan, Yaliang Li et al.
ICML 2024poster
Ranking-based Client Imitation Selection for Efficient Federated Learning
Chunlin Tian, Zhan Shi, Xinpeng Qin et al.
ICML 2024poster
Sparse-IFT: Sparse Iso-FLOP Transformations for Maximizing Training Efficiency
Vithursan Thangarasa, Shreyas Saxena, Abhay Gupta et al.
ICML 2024poster