ICLR 2025 "vision transformers" Papers
11 papers found
A Theoretical Analysis of Self-Supervised Learning for Vision Transformers
Yu Huang, Zixin Wen, Yuejie Chi et al.
ICLR 2025posterarXiv:2403.02233
3
citations
Brain Mapping with Dense Features: Grounding Cortical Semantic Selectivity in Natural Images With Vision Transformers
Andrew Luo, Jacob Yeung, Rushikesh Zawar et al.
ICLR 2025posterarXiv:2410.05266
12
citations
Discovering Influential Neuron Path in Vision Transformers
Yifan Wang, Yifei Liu, Yingdong Shi et al.
ICLR 2025posterarXiv:2503.09046
4
citations
FLOPS: Forward Learning with OPtimal Sampling
Tao Ren, Zishi Zhang, Jinyang Jiang et al.
ICLR 2025posterarXiv:2410.05966
2
citations
LevAttention: Time, Space and Streaming Efficient Algorithm for Heavy Attentions
Ravindran Kannan, Chiranjib Bhattacharyya, Praneeth Kacham et al.
ICLR 2025posterarXiv:2410.05462
1
citations
Locality Alignment Improves Vision-Language Models
Ian Covert, Tony Sun, James Y Zou et al.
ICLR 2025posterarXiv:2410.11087
Metric-Driven Attributions for Vision Transformers
Chase Walker, Sumit Jha, Rickard Ewetz
ICLR 2025poster
1
citations
Morphing Tokens Draw Strong Masked Image Models
Taekyung Kim, Byeongho Heo, Dongyoon Han
ICLR 2025posterarXiv:2401.00254
3
citations
PolaFormer: Polarity-aware Linear Attention for Vision Transformers
Weikang Meng, Yadan Luo, Xin Li et al.
ICLR 2025posterarXiv:2501.15061
36
citations
Spiking Vision Transformer with Saccadic Attention
Shuai Wang, Malu Zhang, Dehao Zhang et al.
ICLR 2025oralarXiv:2502.12677
15
citations
Synergy and Diversity in CLIP: Enhancing Performance Through Adaptive Backbone Ensembling
Cristian Rodriguez-Opazo, Ehsan Abbasnejad, Damien Teney et al.
ICLR 2025posterarXiv:2405.17139
1
citations