"model distillation" Papers
9 papers found
FADA: Fast Diffusion Avatar Synthesis with Mixed-Supervised Multi-CFG Distillation
Tianyun Zhong, Chao Liang, Jianwen Jiang et al.
CVPR 2025posterarXiv:2412.16915
5
citations
Position: Require Frontier AI Labs To Release Small "Analog" Models
Shriyash Upadhyay, Philip Quirke, Narmeen Oozeer et al.
NeurIPS 2025poster
Prismatic Synthesis: Gradient-based Data Diversification Boosts Generalization in LLM Reasoning
Jaehun Jung, Seungju Han, Ximing Lu et al.
NeurIPS 2025spotlightarXiv:2505.20161
15
citations
SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction
Ling Yang, Zhaochen Yu, Tianjun Zhang et al.
ICLR 2025posterarXiv:2410.09008
12
citations
Towards Thinking-Optimal Scaling of Test-Time Compute for LLM Reasoning
Wenkai Yang, Shuming Ma, Yankai Lin et al.
NeurIPS 2025posterarXiv:2502.18080
96
citations
AMD: Automatic Multi-step Distillation of Large-scale Vision Models
Cheng Han, Qifan Wang, Sohail A Dianat et al.
ECCV 2024posterarXiv:2407.04208
14
citations
Knowledge Transfer from Vision Foundation Models for Efficient Training of Small Task-specific Models
Raviteja Vemulapalli, Hadi Pouransari, Fartash Faghri et al.
ICML 2024poster
MGit: A Model Versioning and Management System
Wei Hao, Daniel Mendoza, Rafael Mendes et al.
ICML 2024poster
USTAD: Unified Single-model Training Achieving Diverse Scores for Information Retrieval
Seungyeon Kim, Ankit Singh Rawat, Manzil Zaheer et al.
ICML 2024poster