Zhenheng Tang
5
Papers
65
Total Citations
Papers (5)
STBLLM: Breaking the 1-Bit Barrier with Structured Binary LLMs
ICLR 2025
31
citations
ChunkKV: Semantic-Preserving KV Cache Compression for Efficient Long-Context LLM Inference
NeurIPS 2025
14
citations
The Lottery LLM Hypothesis, Rethinking What Abilities Should LLM Compression Preserve?
ICLR 2025
10
citations
ParZC: Parametric Zero-Cost Proxies for Efficient NAS
AAAI 2025
10
citations
Pruner-Zero: Evolving Symbolic Pruning Metric From Scratch for Large Language Models
ICML 2024
0
citations