2025 Poster "attention mechanisms" Papers
9 papers found
Balanced Token Pruning: Accelerating Vision Language Models Beyond Local Optimization
kaiyuan Li, Xiaoyue Chen, Chen Gao et al.
NEURIPS 2025posterarXiv:2505.22038
4
citations
Beyond Text-Visual Attention: Exploiting Visual Cues for Effective Token Pruning in VLMs
Qizhe Zhang, Aosong Cheng, Ming Lu et al.
ICCV 2025posterarXiv:2412.01818
37
citations
DISCO: Disentangled Communication Steering for Large Language Models
Max Torop, Aria Masoomi, Masih Eskandar et al.
NEURIPS 2025posterarXiv:2509.16820
Fast attention mechanisms: a tale of parallelism
Jingwen Liu, Hantao Yu, Clayton Sanford et al.
NEURIPS 2025posterarXiv:2509.09001
Making Text Embedders Few-Shot Learners
Chaofan Li, Minghao Qin, Shitao Xiao et al.
ICLR 2025posterarXiv:2409.15700
86
citations
On the Role of Attention Heads in Large Language Model Safety
Zhenhong Zhou, Haiyang Yu, Xinghua Zhang et al.
ICLR 2025posterarXiv:2410.13708
40
citations
Rope to Nope and Back Again: A New Hybrid Attention Strategy
Bowen Yang, Bharat Venkitesh, Dwaraknath Gnaneshwar Talupuru et al.
NEURIPS 2025posterarXiv:2501.18795
20
citations
Scale-invariant attention
Ben Anson, Xi Wang, Laurence Aitchison
NEURIPS 2025posterarXiv:2505.17083
2
citations
Why Does the Effective Context Length of LLMs Fall Short?
Chenxin An, Jun Zhang, Ming Zhong et al.
ICLR 2025posterarXiv:2410.18745