ICLR 2025 "model compression" Papers
10 papers found
Composable Interventions for Language Models
Arinbjörn Kolbeinsson, Kyle O'Brien, Tianjin Huang et al.
ICLR 2025posterarXiv:2407.06483
4
citations
HarmAug: Effective Data Augmentation for Knowledge Distillation of Safety Guard Models
Seanie Lee, Haebin Seong, Dong Bok Lee et al.
ICLR 2025posterarXiv:2410.01524
13
citations
LLaMaFlex: Many-in-one LLMs via Generalized Pruning and Weight Sharing
Ruisi Cai, Saurav Muralidharan, Hongxu Yin et al.
ICLR 2025poster
4
citations
Mixture Compressor for Mixture-of-Experts LLMs Gains More
Wei Huang, Yue Liao, Jianhui Liu et al.
ICLR 2025posterarXiv:2410.06270
22
citations
Optimal Brain Apoptosis
Mingyuan Sun, Zheng Fang, Jiaxu Wang et al.
ICLR 2025posterarXiv:2502.17941
3
citations
Quantized Spike-driven Transformer
Xuerui Qiu, Malu Zhang, Jieyuan Zhang et al.
ICLR 2025posterarXiv:2501.13492
14
citations
SeedLM: Compressing LLM Weights into Seeds of Pseudo-Random Generators
Rasoul Shafipour, David Harrison, Maxwell Horton et al.
ICLR 2025posterarXiv:2410.10714
2
citations
SpikeLLM: Scaling up Spiking Neural Network to Large Language Models via Saliency-based Spiking
Xingrun Xing, Boyan Gao, Zheng Liu et al.
ICLR 2025posterarXiv:2407.04752
21
citations
Systematic Outliers in Large Language Models
Yongqi An, Xu Zhao, Tao Yu et al.
ICLR 2025posterarXiv:2502.06415
15
citations
The Unreasonable Ineffectiveness of the Deeper Layers
Andrey Gromov, Kushal Tirumala, Hassan Shapourian et al.
ICLR 2025posterarXiv:2403.17887
160
citations