Poster Papers

24,624 papers found • Page 416 of 493

Mismatch Quest: Visual and Textual Feedback for Image-Text Misalignment

Brian Gordon, Yonatan Bitton, Yonatan Shafir et al.

ECCV 2024arXiv:2312.03766
17
citations

Missing Modality Prediction for Unpaired Multimodal Learning via Joint Embedding of Unimodal Models

Taesup Kim, Donggeun Kim

ECCV 2024arXiv:2407.12616
13
citations

Mitigating Background Shift in Class-Incremental Semantic Segmentation

gilhan Park, WonJun Moon, SuBeen Lee et al.

ECCV 2024arXiv:2407.11859
12
citations

Mitigating Catastrophic Forgetting in Online Continual Learning by Modeling Previous Task Interrelations via Pareto Optimization

Yichen WU, Hong Wang, Peilin Zhao et al.

ICML 2024

Mitigating Emergent Robustness Degradation while Scaling Graph Learning

Xiangchi Yuan, Chunhui Zhang, Yijun Tian et al.

ICLR 2024

Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning

Fuxiao Liu, Kevin Lin, Linjie Li et al.

ICLR 2024arXiv:2306.14565
422
citations

Mitigating Label Noise on Graphs via Topological Sample Selection

Yuhao Wu, Jiangchao Yao, Xiaobo Xia et al.

ICML 2024arXiv:2403.01942
7
citations

Mitigating Motion Blur in Neural Radiance Fields with Events and Frames

Marco Cannici, Davide Scaramuzza

CVPR 2024arXiv:2403.19780
38
citations

Mitigating Noisy Correspondence by Geometrical Structure Consistency Learning

Zihua Zhao, Mengxi Chen, Tianjie Dai et al.

CVPR 2024arXiv:2405.16996
12
citations

Mitigating Object Dependencies: Improving Point Cloud Self-Supervised Learning through Object Exchange

Yanhao Wu, Tong Zhang, Wei Ke et al.

CVPR 2024arXiv:2404.07504
7
citations

Mitigating Oversmoothing Through Reverse Process of GNNs for Heterophilic Graphs

MoonJeong Park, Jaeseung Heo, Dongwoo Kim

ICML 2024arXiv:2403.10543
6
citations

Mitigating Perspective Distortion-induced Shape Ambiguity in Image Crops

Aditya Prakash, Arjun Gupta, Saurabh Gupta

ECCV 2024arXiv:2312.06594
4
citations

Mitigating Privacy Risk in Membership Inference by Convex-Concave Loss

Zhenlong Liu, Lei Feng, HUIPING ZHUANG et al.

ICML 2024arXiv:2402.05453
5
citations

Mitigating the Curse of Dimensionality for Certified Robustness via Dual Randomized Smoothing

Song Xia, Yi Yu, Jiang Xudong et al.

ICLR 2024arXiv:2404.09586
15
citations

MixDQ: Memory-Efficient Few-Step Text-to-Image Diffusion Models with Metric-Decoupled Mixed Precision Quantization

Zhao Tianchen, Xuefei Ning, Tongcheng Fang et al.

ECCV 2024arXiv:2405.17873
37
citations

Mixed-Precision Quantization for Federated Learning on Resource-Constrained Heterogeneous Devices

Huancheng Chen, Haris Vikalo

CVPR 2024arXiv:2311.18129
20
citations

Mixed-Type Tabular Data Synthesis with Score-based Diffusion in Latent Space

Hengrui Zhang, Jiani Zhang, Zhengyuan Shen et al.

ICLR 2024arXiv:2310.09656
199
citations

MixSATGEN: Learning Graph Mixing for SAT Instance Generation

Xinyan Chen, Yang Li, Runzhong Wang et al.

ICLR 2024

MixSup: Mixed-grained Supervision for Label-efficient LiDAR-based 3D Object Detection

Yuxue Yang, Lue Fan, Zhaoxiang Zhang

ICLR 2024arXiv:2401.16305
7
citations

Mixture of Efficient Diffusion Experts Through Automatic Interval and Sub-Network Selection

Alireza Ganjdanesh, Yan Kang, Yuchen Liu et al.

ECCV 2024arXiv:2409.15557
12
citations

Mixture-of-Experts Meets Instruction Tuning: A Winning Combination for Large Language Models

Sheng Shen, Le Hou, Yanqi Zhou et al.

ICLR 2024

Mixture of LoRA Experts

xun wu, Shaohan Huang, Furu Wei

ICLR 2024arXiv:2404.13628
136
citations

Mixture of Weak and Strong Experts on Graphs

Hanqing Zeng, Hanjia Lyu, Diyi Hu et al.

ICLR 2024
10
citations

MLAgentBench: Evaluating Language Agents on Machine Learning Experimentation

Qian Huang, Jian Vora, Percy Liang et al.

ICML 2024arXiv:2310.03302
168
citations

MLI Formula: A Nearly Scale-Invariant Solution with Noise Perturbation

Bowen Tao, Xin-Chun Li, De-Chuan Zhan

ICML 2024

MLIP: Efficient Multi-Perspective Language-Image Pretraining with Exhaustive Data Utilization

Yu Zhang, Qi Zhang, Zixuan Gong et al.

ICML 2024arXiv:2406.01460
7
citations

MLIP: Enhancing Medical Visual Representation with Divergence Encoder and Knowledge-guided Contrastive Learning

Zhe Li, Laurence Yang, Bocheng Ren et al.

CVPR 2024arXiv:2402.02045
34
citations

MLLM-as-a-Judge: Assessing Multimodal LLM-as-a-Judge with Vision-Language Benchmark

Dongping Chen, Ruoxi Chen, Shilin Zhang et al.

ICML 2024arXiv:2402.04788
281
citations

MLP Can Be A Good Transformer Learner

Sihao Lin, Pumeng Lyu, Dongrui Liu et al.

CVPR 2024arXiv:2404.05657
21
citations

MLPHand: Real Time Multi-View 3D Hand Reconstruction via MLP Modeling

Jian Yang, Jiakun Li, Guoming Li et al.

ECCV 2024

ML-SemReg: Boosting Point Cloud Registration with Multi-level Semantic Consistency

Shaocheng Yan, Pengcheng Shi, Jiayuan Li

ECCV 2024arXiv:2407.09862
6
citations

MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training

Brandon McKinzie, Zhe Gan, Jean-Philippe Fauconnier et al.

ECCV 2024arXiv:2403.09611
250
citations

MMA-Diffusion: MultiModal Attack on Diffusion Models

Yijun Yang, Ruiyuan Gao, Xiaosen Wang et al.

CVPR 2024arXiv:2311.17516
150
citations

MMA: Multi-Modal Adapter for Vision-Language Models

Lingxiao Yang, Ru-Yuan Zhang, Yanchen Wang et al.

CVPR 2024

MMBENCH: Is Your Multi-Modal Model an All-around Player?

Yuan Liu, Haodong Duan, Yuanhan Zhang et al.

ECCV 2024arXiv:2307.06281
1745
citations

MMCert: Provable Defense against Adversarial Attacks to Multi-modal Models

Yanting Wang, Hongye Fu, Wei Zou et al.

CVPR 2024arXiv:2403.19080
5
citations

MMEarth: Exploring Multi-Modal Pretext Tasks For Geospatial Representation Learning

Vishal Nedungadi, Ankit Kariryaa, Stefan Oehmcke et al.

ECCV 2024arXiv:2405.02771
68
citations

MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning

Haozhe Zhao, Zefan Cai, Shuzheng Si et al.

ICLR 2024arXiv:2309.07915
191
citations

MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI

Xiang Yue, Yuansheng Ni, Kai Zhang et al.

CVPR 2024arXiv:2311.16502
1715
citations

MMPareto: Boosting Multimodal Learning with Innocent Unimodal Assistance

Yake Wei, Di Hu

ICML 2024arXiv:2405.17730
64
citations

m&m’s: A Benchmark to Evaluate Tool-Use for multi-step multi-modal Tasks

Zixian Ma, Weikai Huang, Jieyu Zhang et al.

ECCV 2024arXiv:2403.11085
36
citations

MM-SafetyBench: A Benchmark for Safety Evaluation of Multimodal Large Language Models

Xin Liu, Yichen Zhu, Jindong Gu et al.

ECCV 2024arXiv:2311.17600
199
citations

MMT-Bench: A Comprehensive Multimodal Benchmark for Evaluating Large Vision-Language Models Towards Multitask AGI

Kaining Ying, Fanqing Meng, Jin Wang et al.

ICML 2024arXiv:2404.16006
163
citations

MM-Vet: Evaluating Large Multimodal Models for Integrated Capabilities

Weihao Yu, Zhengyuan Yang, Linjie Li et al.

ICML 2024arXiv:2308.02490
1066
citations

MMVP: A Multimodal MoCap Dataset with Vision and Pressure Sensors

He Zhang, Shenghao Ren, Haolei Yuan et al.

CVPR 2024arXiv:2403.17610
17
citations

MMVR: Millimeter-wave Multi-View Radar Dataset and Benchmark for Indoor Perception

Mohammad Mahbubur Rahman, Ryoma Yataka, Sorachi Kato et al.

ECCV 2024arXiv:2406.10708
20
citations

MoAI: Mixture of All Intelligence for Large Language and Vision Models

Byung-Kwan Lee, Beomchan Park, Chae Won Kim et al.

ECCV 2024arXiv:2403.07508
34
citations

Mobile Attention: Mobile-Friendly Linear-Attention for Vision Transformers

Zhiyu Yao, Jian Wang, Haixu Wu et al.

ICML 2024

MobileCLIP: Fast Image-Text Models through Multi-Modal Reinforced Training

Pavan Kumar Anasosalu Vasu, Hadi Pouransari, Fartash Faghri et al.

CVPR 2024arXiv:2311.17049
90
citations

MobileDiffusion: Instant Text-to-Image Generation on Mobile Devices

Yang Zhao, Zhisheng Xiao, Yanwu Xu et al.

ECCV 2024arXiv:2311.16567
36
citations