"attention mechanism efficiency" Papers
2 papers found
EDiT: Efficient Diffusion Transformers with Linear Compressed Attention
Philipp Becker, Abhinav Mehrotra, Ruchika Chavhan et al.
ICCV 2025posterarXiv:2503.16726
3
citations
Superposition Prompting: Improving and Accelerating Retrieval-Augmented Generation
Thomas Merth, Qichen Fu, Mohammad Rastegari et al.
ICML 2024poster