NeurIPS 2025 "in-context learning" Papers
13 papers found
Attention-based clustering
Rodrigo Maulen Soto, Pierre Marion, Claire Boyer
NeurIPS 2025posterarXiv:2505.13112
Bridging Sign and Spoken Languages: Pseudo Gloss Generation for Sign Language Translation
Jianyuan Guo, Peike Li, Trevor Cohn
NeurIPS 2025oralarXiv:2505.15438
3
citations
Explore In-Context Message Passing Operator for Graph Neural Networks in A Mean Field Game
Tingting Dan, Xinwei Huang, Won Hwa Kim et al.
NeurIPS 2025poster
Knowledge Starts with Practice: Knowledge-Aware Exercise Generative Recommendation with Adaptive Multi-Agent Cooperation
Yangtao Zhou, Hua Chu, chen et al.
NeurIPS 2025poster
Learning to Rank for In-Context Example Retrieval
Yuwen Ji, Luodan Zhang, Ambyer han et al.
NeurIPS 2025poster
Optimal Dynamic Regret by Transformers for Non-Stationary Reinforcement Learning
Baiyuan Chen, Shinji Ito, Masaaki Imaizumi
NeurIPS 2025posterarXiv:2508.16027
Reasoning Models Better Express Their Confidence
Dongkeun Yoon, Seungone Kim, Sohee Yang et al.
NeurIPS 2025posterarXiv:2505.14489
32
citations
Self-Generated In-Context Examples Improve LLM Agents for Sequential Decision-Making Tasks
Vishnu Sarukkai, Zhiqiang Xie, Kayvon Fatahalian
NeurIPS 2025posterarXiv:2505.00234
4
citations
Theoretical Insights into In-context Learning with Unlabeled Data
Yingcong Li, Xiangyu Chang, Muti Kara et al.
NeurIPS 2025poster
Transformers are almost optimal metalearners for linear classification
Roey Magen, Gal Vardi
NeurIPS 2025posterarXiv:2510.19797
1
citations
Unlabeled Data Can Provably Enhance In-Context Learning of Transformers
Renpu Liu, Jing Yang
NeurIPS 2025posterarXiv:2601.10058
1
citations
Vision-centric Token Compression in Large Language Model
Ling Xing, Alex Jinpeng Wang, Rui Yan et al.
NeurIPS 2025spotlightarXiv:2502.00791
7
citations
What One Cannot, Two Can: Two-Layer Transformers Provably Represent Induction Heads on Any-Order Markov Chains
Chanakya Ekbote, Ashok Vardhan Makkuva, Marco Bondaschi et al.
NeurIPS 2025spotlightarXiv:2508.07208