2025 "multi-modal llms" Papers
7 papers found
$F^3Set$: Towards Analyzing Fast, Frequent, and Fine-grained Events from Videos
Zhaoyu Liu, Kan Jiang, Murong Ma et al.
ICLR 2025oral
3
citations
AC-LoRA: (Almost) Training-Free Access Control Aware Multi-Modal LLMs
Lara Magdalena Lazier, Aritra Dhar, Vasilije Stambolic et al.
NEURIPS 2025posterarXiv:2505.11557
1
citations
AIM: Adaptive Inference of Multi-Modal LLMs via Token Merging and Pruning
Yiwu Zhong, Zhuoming Liu, Yin Li et al.
ICCV 2025posterarXiv:2412.03248
21
citations
Can Multi-Modal LLMs Provide Live Step-by-Step Task Guidance?
Apratim Bhattacharyya, Bicheng Xu, Sanjay Haresh et al.
NEURIPS 2025posterarXiv:2511.21998
Distilling Multi-modal Large Language Models for Autonomous Driving
Deepti Hegde, Rajeev Yasarla, Hong Cai et al.
CVPR 2025posterarXiv:2501.09757
27
citations
The Labyrinth of Links: Navigating the Associative Maze of Multi-modal LLMs
HONG LI, Nanxi Li, Yuanjie Chen et al.
ICLR 2025posterarXiv:2410.01417
3
citations
Video-MME: The First-Ever Comprehensive Evaluation Benchmark of Multi-modal LLMs in Video Analysis
Chaoyou Fu, Yuhan Dai, Yongdong Luo et al.
CVPR 2025highlightarXiv:2405.21075
876
citations