ICLR 2025 "transformer models" Papers
7 papers found
Can In-context Learning Really Generalize to Out-of-distribution Tasks?
Qixun Wang, Yifei Wang, Xianghua Ying et al.
ICLR 2025posterarXiv:2410.09695
15
citations
Learning Randomized Algorithms with Transformers
Johannes von Oswald, Seijin Kobayashi, Yassir Akram et al.
ICLR 2025posterarXiv:2408.10818
1
citations
Mixture of Parrots: Experts improve memorization more than reasoning
Samy Jelassi, Clara Mohri, David Brandfonbrener et al.
ICLR 2025posterarXiv:2410.19034
14
citations
Multi-modal brain encoding models for multi-modal stimuli
SUBBA REDDY OOTA, Khushbu Pahwa, mounika marreddy et al.
ICLR 2025posterarXiv:2505.20027
9
citations
SelectFormer in Data Markets: Privacy-Preserving and Efficient Data Selection for Transformers with Multi-Party Computation
Xu Ouyang, Felix Xiaozhu Lin, Yangfeng Ji
ICLR 2025poster
Toward Understanding In-context vs. In-weight Learning
Bryan Chan, Xinyi Chen, Andras Gyorgy et al.
ICLR 2025posterarXiv:2410.23042
14
citations
Tracing Representation Progression: Analyzing and Enhancing Layer-Wise Similarity
Jiachen Jiang, Jinxin Zhou, Zhihui Zhu
ICLR 2025posterarXiv:2406.14479
16
citations