2025 "self-distillation" Papers

15 papers found

Beyond Autoregression: Fast LLMs via Self-Distillation Through Time

Justin Deschenaux, Caglar Gulcehre

ICLR 2025posterarXiv:2410.21035
25
citations

COIN: Confidence Score-Guided Distillation for Annotation-Free Cell Segmentation

Sanghyun Jo, Seo Lee, Seungwoo Lee et al.

ICCV 2025posterarXiv:2503.11439
2
citations

DebGCD: Debiased Learning with Distribution Guidance for Generalized Category Discovery

Yuanpei Liu, Kai Han

ICLR 2025posterarXiv:2504.04804
7
citations

FedLPA: Local Prior Alignment for Heterogeneous Federated Generalized Category Discovery

Geeho Kim, Jinu Lee, Bohyung Han

NEURIPS 2025poster

How to build a consistency model: Learning flow maps via self-distillation

Nicholas Boffi, Michael Albergo, Eric Vanden-Eijnden

NEURIPS 2025posterarXiv:2505.18825
25
citations

LoftUp: Learning a Coordinate-Based Feature Upsampler for Vision Foundation Models

Haiwen Huang, Anpei Chen, Volodymyr Havrylov et al.

ICCV 2025posterarXiv:2504.14032
10
citations

Long-tailed Adversarial Training with Self-Distillation

Seungju Cho, Hongsin Lee, Changick Kim

ICLR 2025posterarXiv:2503.06461
1
citations

MAESTRO: Masked Encoding Set Transformer with Self-Distillation

Matthew Lee, Jaesik Kim, Matei Ionita et al.

ICLR 2025poster

Rethinking Self-Distillation: Label Averaging and Enhanced Soft Label Refinement with Partial Labels

Hyeonsu Jeong, Hye Won Chung

ICLR 2025posterarXiv:2402.10482
1
citations

SeerAttention: Self-distilled Attention Gating for Efficient Long-context Prefilling

Yizhao Gao, Zhichen Zeng, DaYou Du et al.

NEURIPS 2025poster

Shortcutting Pre-trained Flow Matching Diffusion Models is Almost Free Lunch

Xu Cai, Yang Wu, Qianli Chen et al.

NEURIPS 2025posterarXiv:2510.17858

Synergy Between the Strong and the Weak: Spiking Neural Networks are Inherently Self-Distillers

Yongqi Ding, Lin Zuo, Mengmeng Jing et al.

NEURIPS 2025oralarXiv:2510.07924

The Effect of Optimal Self-Distillation in Noisy Gaussian Mixture Model

Kaito Takanami, Takashi Takahashi, Ayaka Sakata

NEURIPS 2025posterarXiv:2501.16226
2
citations

Unbiased Region-Language Alignment for Open-Vocabulary Dense Prediction

Yunheng Li, Yuxuan Li, Quan-Sheng Zeng et al.

ICCV 2025posterarXiv:2412.06244
6
citations

Vision Transformers with Self-Distilled Registers

Zipeng Yan, Yinjie Chen, Chong Zhou et al.

NEURIPS 2025spotlightarXiv:2505.21501
4
citations