ICLR Oral "knowledge distillation" Papers
2 papers found
SLMRec: Distilling Large Language Models into Small for Sequential Recommendation
Wujiang Xu, Qitian Wu, Zujie Liang et al.
ICLR 2025oralarXiv:2405.17890
17
citations
TAID: Temporally Adaptive Interpolated Distillation for Efficient Knowledge Transfer in Language Models
Makoto Shing, Kou Misaki, Han Bao et al.
ICLR 2025oralarXiv:2501.16937
12
citations