Taesu Kim
3
Papers
101
Total Citations
Papers (3)
OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Models
AAAI 2024arXiv
100
citations
GraLoRA: Granular Low-Rank Adaptation for Parameter-Efficient Fine-Tuning
NeurIPS 2025arXiv
1
citations
SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks
ICML 2024
0
citations