2025 Spotlight "in-context learning" Papers
2 papers found
Vision-centric Token Compression in Large Language Model
Ling Xing, Alex Jinpeng Wang, Rui Yan et al.
NeurIPS 2025spotlightarXiv:2502.00791
7
citations
What One Cannot, Two Can: Two-Layer Transformers Provably Represent Induction Heads on Any-Order Markov Chains
Chanakya Ekbote, Ashok Vardhan Makkuva, Marco Bondaschi et al.
NeurIPS 2025spotlightarXiv:2508.07208