"multi-modal fusion" Papers
7 papers found
Grounding 3D Object Affordance with Language Instructions, Visual Observations and Interactions
He Zhu, Quyu Kong, Kechun Xu et al.
CVPR 2025posterarXiv:2504.04744
6
citations
Tri-MARF: A Tri-Modal Multi-Agent Responsive Framework for Comprehensive 3D Object Annotation
jusheng zhang, Yijia Fan, Zimo Wen et al.
NeurIPS 2025poster
Adaptive Multi-modal Fusion of Spatially Variant Kernel Refinement with Diffusion Model for Blind Image Super-Resolution
Junxiong Lin, Yan Wang, Zeng Tao et al.
ECCV 2024posterarXiv:2403.05808
5
citations
A Multi-Modal Contrastive Diffusion Model for Therapeutic Peptide Generation
Yongkang Wang, Xuan Liu, Feng Huang et al.
AAAI 2024paperarXiv:2312.15665
22
citations
Beyond the Label Itself: Latent Labels Enhance Semi-supervised Point Cloud Panoptic Segmentation
Yujun Chen, Xin Tan, Zhizhong Zhang et al.
AAAI 2024paperarXiv:2312.08234
6
citations
Bridging the Gap between 2D and 3D Visual Question Answering: A Fusion Approach for 3D VQA
Wentao Mo, Yang Liu
AAAI 2024paperarXiv:2402.15933
26
citations
DrFuse: Learning Disentangled Representation for Clinical Multi-Modal Fusion with Missing Modality and Modal Inconsistency
Wenfang Yao, Kejing Yin, William Cheung et al.
AAAI 2024paperarXiv:2403.06197
55
citations