Paper "model compression" Papers
10 papers found
An Empirical Study of CLIP for Text-Based Person Search
Cao Min, Yang Bai, ziyin Zeng et al.
AAAI 2024paperarXiv:2308.10045
94
citations
BiPFT: Binary Pre-trained Foundation Transformer with Low-Rank Estimation of Binarization Residual Polynomials
Xingrun Xing, Li Du, Xinyuan Wang et al.
AAAI 2024paperarXiv:2312.08937
Building Variable-Sized Models via Learngene Pool
Boyu Shi, Shiyu Xia, Xu Yang et al.
AAAI 2024paperarXiv:2312.05743
5
citations
Entropy Induced Pruning Framework for Convolutional Neural Networks
Yiheng Lu, Ziyu Guan, Yaming Yang et al.
AAAI 2024paperarXiv:2208.06660
6
citations
EPSD: Early Pruning with Self-Distillation for Efficient Model Compression
Dong Chen, Ning Liu, Yichen Zhu et al.
AAAI 2024paperarXiv:2402.00084
8
citations
Exploring Post-training Quantization in LLMs from Comprehensive Study to Low Rank Compensation
Zhewei Yao, Xiaoxia Wu, Cheng Li et al.
AAAI 2024paperarXiv:2303.08302
70
citations
Fluctuation-Based Adaptive Structured Pruning for Large Language Models
Yongqi An, Xu Zhao, Tao Yu et al.
AAAI 2024paperarXiv:2312.11983
96
citations
Generative Model-Based Feature Knowledge Distillation for Action Recognition
Guiqin Wang, Peng Zhao, Yanjiang Shi et al.
AAAI 2024paperarXiv:2312.08644
6
citations
OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Models
Changhun Lee, Jungyu Jin, Taesu Kim et al.
AAAI 2024paperarXiv:2306.02272
100
citations
Progressive Distillation Based on Masked Generation Feature Method for Knowledge Graph Completion
Cunhang Fan, Yujie Chen, Jun Xue et al.
AAAI 2024paperarXiv:2401.12997
5
citations