2025 "self-distillation" Papers
15 papers found
Beyond Autoregression: Fast LLMs via Self-Distillation Through Time
Justin Deschenaux, Caglar Gulcehre
ICLR 2025posterarXiv:2410.21035
25
citations
COIN: Confidence Score-Guided Distillation for Annotation-Free Cell Segmentation
Sanghyun Jo, Seo Lee, Seungwoo Lee et al.
ICCV 2025posterarXiv:2503.11439
2
citations
DebGCD: Debiased Learning with Distribution Guidance for Generalized Category Discovery
Yuanpei Liu, Kai Han
ICLR 2025posterarXiv:2504.04804
7
citations
FedLPA: Local Prior Alignment for Heterogeneous Federated Generalized Category Discovery
Geeho Kim, Jinu Lee, Bohyung Han
NEURIPS 2025poster
How to build a consistency model: Learning flow maps via self-distillation
Nicholas Boffi, Michael Albergo, Eric Vanden-Eijnden
NEURIPS 2025posterarXiv:2505.18825
25
citations
LoftUp: Learning a Coordinate-Based Feature Upsampler for Vision Foundation Models
Haiwen Huang, Anpei Chen, Volodymyr Havrylov et al.
ICCV 2025posterarXiv:2504.14032
10
citations
Long-tailed Adversarial Training with Self-Distillation
Seungju Cho, Hongsin Lee, Changick Kim
ICLR 2025posterarXiv:2503.06461
1
citations
MAESTRO: Masked Encoding Set Transformer with Self-Distillation
Matthew Lee, Jaesik Kim, Matei Ionita et al.
ICLR 2025poster
Rethinking Self-Distillation: Label Averaging and Enhanced Soft Label Refinement with Partial Labels
Hyeonsu Jeong, Hye Won Chung
ICLR 2025posterarXiv:2402.10482
1
citations
SeerAttention: Self-distilled Attention Gating for Efficient Long-context Prefilling
Yizhao Gao, Zhichen Zeng, DaYou Du et al.
NEURIPS 2025poster
Shortcutting Pre-trained Flow Matching Diffusion Models is Almost Free Lunch
Xu Cai, Yang Wu, Qianli Chen et al.
NEURIPS 2025posterarXiv:2510.17858
Synergy Between the Strong and the Weak: Spiking Neural Networks are Inherently Self-Distillers
Yongqi Ding, Lin Zuo, Mengmeng Jing et al.
NEURIPS 2025oralarXiv:2510.07924
The Effect of Optimal Self-Distillation in Noisy Gaussian Mixture Model
Kaito Takanami, Takashi Takahashi, Ayaka Sakata
NEURIPS 2025posterarXiv:2501.16226
2
citations
Unbiased Region-Language Alignment for Open-Vocabulary Dense Prediction
Yunheng Li, Yuxuan Li, Quan-Sheng Zeng et al.
ICCV 2025posterarXiv:2412.06244
6
citations
Vision Transformers with Self-Distilled Registers
Zipeng Yan, Yinjie Chen, Chong Zhou et al.
NEURIPS 2025spotlightarXiv:2505.21501
4
citations