NeurIPS "attention mechanisms" Papers
6 papers found
Balanced Token Pruning: Accelerating Vision Language Models Beyond Local Optimization
kaiyuan Li, Xiaoyue Chen, Chen Gao et al.
NeurIPS 2025posterarXiv:2505.22038
4
citations
DISCO: Disentangled Communication Steering for Large Language Models
Max Torop, Aria Masoomi, Masih Eskandar et al.
NeurIPS 2025posterarXiv:2509.16820
Fast attention mechanisms: a tale of parallelism
Jingwen Liu, Hantao Yu, Clayton Sanford et al.
NeurIPS 2025posterarXiv:2509.09001
Neural Fractional Attention Differential Equations
Qiyu Kang, Wenjun Cui, Xuhao Li et al.
NeurIPS 2025oral
Scale-invariant attention
Ben Anson, Xi Wang, Laurence Aitchison
NeurIPS 2025posterarXiv:2505.17083
2
citations
ZeroS: Zero‑Sum Linear Attention for Efficient Transformers
Jiecheng Lu, Xu Han, Yan Sun et al.
NeurIPS 2025spotlight