2025 Poster "linear attention" Papers
6 papers found
Alias-Free ViT: Fractional Shift Invariance via Linear Attention
Hagay Michaeli, Daniel Soudry
NeurIPS 2025posterarXiv:2510.22673
Degrees of Freedom for Linear Attention: Distilling Softmax Attention with Optimal Feature Efficiency
Naoki Nishikawa, Rei Higuchi, Taiji Suzuki
NeurIPS 2025posterarXiv:2507.03340
1
citations
Jet-Nemotron: Efficient Language Model with Post Neural Architecture Search
Yuxian Gu, Qinghao Hu, Haocheng Xi et al.
NeurIPS 2025posterarXiv:2508.15884
15
citations
Physics of Language Models: Part 4.1, Architecture Design and the Magic of Canon Layers
Zeyuan Allen-Zhu
NeurIPS 2025posterarXiv:2512.17351
8
citations
PolaFormer: Polarity-aware Linear Attention for Vision Transformers
Weikang Meng, Yadan Luo, Xin Li et al.
ICLR 2025posterarXiv:2501.15061
36
citations
ThunderKittens: Simple, Fast, and $\textit{Adorable}$ Kernels
Benjamin Spector, Simran Arora, Aaryan Singhal et al.
ICLR 2025poster
3
citations