2025 "multimodal in-context learning" Papers
4 papers found
CoMM: A Coherent Interleaved Image-Text Dataset for Multimodal Understanding and Generation
Wei Chen, Lin Li, Yongqi Yang et al.
CVPR 2025highlightarXiv:2406.10462
12
citations
OmniCorpus: A Unified Multimodal Corpus of 10 Billion-Level Images Interleaved with Text
Qingyun Li, Zhe Chen, Weiyun Wang et al.
ICLR 2025posterarXiv:2406.08418
48
citations
SMMILE: An expert-driven benchmark for multimodal medical in-context learning
Melanie Rieff, Maya Varma, Ossian Rabow et al.
NEURIPS 2025posterarXiv:2506.21355
3
citations
VL-ICL Bench: The Devil in the Details of Multimodal In-Context Learning
Yongshuo Zong, Ondrej Bohdal, Timothy Hospedales
ICLR 2025posterarXiv:2403.13164
17
citations