Poster "model compression" Papers
61 papers found • Page 2 of 2
Junk DNA Hypothesis: Pruning Small Pre-Trained Weights $\textit{Irreversibly}$ and $\textit{Monotonically}$ Impairs ``Difficult" Downstream Tasks in LLMs
Lu Yin, Ajay Jaiswal, Shiwei Liu et al.
ICML 2024poster
KernelWarehouse: Rethinking the Design of Dynamic Convolution
Chao Li, Anbang Yao
ICML 2024poster
Lightweight Image Super-Resolution via Flexible Meta Pruning
Yulun Zhang, Kai Zhang, Luc Van Gool et al.
ICML 2024poster
Localizing Task Information for Improved Model Merging and Compression
Ke Wang, Nikolaos Dimitriadis, Guillermo Ortiz-Jimenez et al.
ICML 2024poster
Pruner-Zero: Evolving Symbolic Pruning Metric From Scratch for Large Language Models
Peijie Dong, Lujun Li, Zhenheng Tang et al.
ICML 2024poster
Rethinking Optimization and Architecture for Tiny Language Models
Yehui Tang, Kai Han, Fangcheng Liu et al.
ICML 2024poster
Reweighted Solutions for Weighted Low Rank Approximation
David Woodruff, Taisuke Yasuda
ICML 2024poster
SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks
Jiwon Song, Kyungseok Oh, Taesu Kim et al.
ICML 2024poster
Soft Prompt Recovers Compressed LLMs, Transferably
Zhaozhuo Xu, Zirui Liu, Beidi Chen et al.
ICML 2024poster
Towards efficient deep spiking neural networks construction with spiking activity based pruning
Yaxin Li, Qi Xu, Jiangrong Shen et al.
ICML 2024poster
Transferring Knowledge From Large Foundation Models to Small Downstream Models
Shikai Qiu, Boran Han, Danielle Robinson et al.
ICML 2024poster