ICML Papers

5,975 papers found • Page 38 of 120

Mitigating Heterogeneous Token Overfitting in LLM Knowledge Editing

Tianci Liu, Ruirui Li, Zihan Dong et al.

ICML 2025posterarXiv:2502.00602

Mitigating Local Cohesion and Global Sparseness in Graph Contrastive Learning with Fuzzy Boundaries

Yuena Lin, Haichun Cai, Jun-Yi Hang et al.

ICML 2025poster

Mitigating Object Hallucination in Large Vision-Language Models via Image-Grounded Guidance

Linxi Zhao, Yihe Deng, Weitong Zhang et al.

ICML 2025spotlightarXiv:2402.08680

MITIGATING OVER-EXPLORATION IN LATENT SPACE OPTIMIZATION USING LES

Omer Ronen, Ahmed Imtiaz Humayun, Richard Baraniuk et al.

ICML 2025posterarXiv:2406.09657

Mitigating Over-Squashing in Graph Neural Networks by Spectrum-Preserving Sparsification

Langzhang Liang, Fanchen Bu, Zixing Song et al.

ICML 2025posterarXiv:2506.16110

Mitigating Plasticity Loss in Continual Reinforcement Learning by Reducing Churn

Hongyao Tang, Johan Obando-Ceron, Pablo Samuel Castro et al.

ICML 2025oralarXiv:2506.00592
3
citations

MixBridge: Heterogeneous Image-to-Image Backdoor Attack through Mixture of Schrödinger Bridges

Shixi Qin, Zhiyong Yang, Shilong Bao et al.

ICML 2025posterarXiv:2505.08809

Mixed-curvature decision trees and random forests

Philippe Chlenski, Quentin Chu, Raiyan Khan et al.

ICML 2025posterarXiv:2406.05227

MixMin: Finding Data Mixtures via Convex Minimization

Anvith Thudi, Evianne Rovers, Yangjun Ruan et al.

ICML 2025posterarXiv:2502.10510

Mixture of Experts Made Intrinsically Interpretable

Xingyi Yang, Constantin Venhoff, Ashkan Khakzar et al.

ICML 2025posterarXiv:2503.07639

Mixture of Experts Provably Detect and Learn the Latent Cluster Structure in Gradient-Based Learning

Ryotaro Kawata, Kohsei Matsutani, Yuri Kinoshita et al.

ICML 2025posterarXiv:2506.01656

Mixture of Hidden-Dimensions: Not All Hidden-States’ Dimensions are Needed in Transformer

Yilong Chen, Junyuan Shang, Zhenyu Zhang et al.

ICML 2025poster

Mixture of Lookup Experts

Shibo Jie, Yehui Tang, Kai Han et al.

ICML 2025oralarXiv:2503.15798

ML$^2$-GCL: Manifold Learning Inspired Lightweight Graph Contrastive Learning

Jianqing Liang, Zhiqiang Li, Xinkai Wei et al.

ICML 2025poster

MME-CoT: Benchmarking Chain-of-Thought in Large Multimodal Models for Reasoning Quality, Robustness, and Efficiency

Dongzhi Jiang, Renrui Zhang, Ziyu Guo et al.

ICML 2025posterarXiv:2502.09621
88
citations

MMedPO: Aligning Medical Vision-Language Models with Clinical-Aware Multimodal Preference Optimization

Kangyu Zhu, Peng Xia, Yun Li et al.

ICML 2025posterarXiv:2412.06141
16
citations

MMInference: Accelerating Pre-filling for Long-Context Visual Language Models via Modality-Aware Permutation Sparse Attention

Yucheng Li, Huiqiang Jiang, Chengruidong Zhang et al.

ICML 2025oral

MM-RLHF: The Next Step Forward in Multimodal LLM Alignment

Yi-Fan Zhang, Tao Yu, Haochen Tian et al.

ICML 2025posterarXiv:2502.10391

Modalities Contribute Unequally: Enhancing Medical Multi-modal Learning through Adaptive Modality Token Re-balancing

Jie Peng, Jenna Ballard, Mohan Zhang et al.

ICML 2025poster

MODA: MOdular Duplex Attention for Multimodal Perception, Cognition, and Emotion Understanding

Zhicheng Zhang, Wuyou Xia, Chenxi Zhao et al.

ICML 2025spotlightarXiv:2507.04635
4
citations

Model-Based Exploration in Monitored Markov Decision Processes

Alireza Kazemipour, Matthew Taylor, Michael Bowling

ICML 2025posterarXiv:2502.16772

Model Immunization from a Condition Number Perspective

Amber Yijia Zheng, Cedar Site Bai, Brian Bullins et al.

ICML 2025oralarXiv:2505.23760

Modeling All-Atom Glycan Structures via Hierarchical Message Passing and Multi-Scale Pre-training

Minghao Xu, Jiaze Song, Keming Wu et al.

ICML 2025posterarXiv:2506.01376

Modeling Multi-Task Model Merging as Adaptive Projective Gradient Descent

Yongxian Wei, Anke Tang, Li Shen et al.

ICML 2025posterarXiv:2501.01230

Models of Heavy-Tailed Mechanistic Universality

Liam Hodgkinson, Zhichao Wang, Michael Mahoney

ICML 2025posterarXiv:2506.03470

Model Steering: Learning with a Reference Model Improves Generalization Bounds and Scaling Laws

Xiyuan Wei, Ming Lin, Fanjiang Ye et al.

ICML 2025spotlightarXiv:2505.06699

Model Swarms: Collaborative Search to Adapt LLM Experts via Swarm Intelligence

Shangbin Feng, Zifeng Wang, Yike Wang et al.

ICML 2025posterarXiv:2410.11163
16
citations

Model Uncertainty Quantification by Conformal Prediction in Continual Learning

Rui Gao, Weiwei Liu

ICML 2025poster
1
citations

Modified K-means Algorithm with Local Optimality Guarantees

Mingyi Li, Michael R. Metel, Akiko Takeda

ICML 2025posterarXiv:2506.06990

Modular Duality in Deep Learning

Jeremy Bernstein, Laker Newhouse

ICML 2025posterarXiv:2410.21265

Modularized Self-Reflected Video Reasoner for Multimodal LLM with Application to Video Question Answering

Zihan Song, Xin Wang, Zi Qian et al.

ICML 2025oral

Modulated Diffusion: Accelerating Generative Modeling with Modulated Quantization

Weizhi Gao, Zhichao Hou, Junqi Yin et al.

ICML 2025posterarXiv:2506.22463

MODULI: Unlocking Preference Generalization via Diffusion Models for Offline Multi-Objective Reinforcement Learning

Yifu Yuan, Zhenrui Zheng, Zibin Dong et al.

ICML 2025posterarXiv:2408.15501

MoEQuant: Enhancing Quantization for Mixture-of-Experts Large Language Models via Expert-Balanced Sampling and Affinity Guidance

Zhixuan Chen, Xing Hu, Dawei Yang et al.

ICML 2025posterarXiv:2505.03804
8
citations

MoE-SVD: Structured Mixture-of-Experts LLMs Compression via Singular Value Decomposition

Wei Li, Lujun Li, Hao Gu et al.

ICML 2025poster

MOGIC: Metadata-infused Oracle Guidance for Improved Extreme Classification

Suchith Chidananda Prabhu, Bhavyajeet Singh, Anshul Mittal et al.

ICML 2025poster

MoHAVE: Mixture of Hierarchical Audio-Visual Experts for Robust Speech Recognition

Sungnyun Kim, Kangwook Jang, Sangmin Bae et al.

ICML 2025posterarXiv:2502.10447

MoH: Multi-Head Attention as Mixture-of-Head Attention

Peng Jin, Bo Zhu, Li Yuan et al.

ICML 2025posterarXiv:2410.11842
37
citations

Moirai-MoE: Empowering Time Series Foundation Models with Sparse Mixture of Experts

Xu Liu, Juncheng Liu, Gerald Woo et al.

ICML 2025posterarXiv:2410.10469

MoMa: Modulating Mamba for Adapting Image Foundation Models to Video Recognition

Yuhuan Yang, Chaofan Ma, Zhenjie Mao et al.

ICML 2025oralarXiv:2506.23283

Momentum-Driven Adaptivity: Towards Tuning-Free Asynchronous Federated Learning

Wenjing Yan, Xiangyu Zhong, Xiaolu Wang et al.

ICML 2025poster

MONA: Myopic Optimization with Non-myopic Approval Can Mitigate Multi-step Reward Hacking

Sebastian Farquhar, Vikrant Varma, David Lindner et al.

ICML 2025posterarXiv:2501.13011

Monte Carlo Tree Diffusion for System 2 Planning

Jaesik Yoon, Hyeonseo Cho, Doojin Baek et al.

ICML 2025spotlightarXiv:2502.07202
17
citations

Monte Carlo Tree Search for Comprehensive Exploration in LLM-Based Automatic Heuristic Design

Zhi Zheng, Zhuoliang Xie, Zhenkun Wang et al.

ICML 2025posterarXiv:2501.08603

Monte-Carlo Tree Search with Uncertainty Propagation via Optimal Transport

Tuan Dam, Pascal Stenger, Lukas Schneider et al.

ICML 2025spotlightarXiv:2309.10737
2
citations

MoRAgent: Parameter Efficient Agent Tuning with Mixture-of-Roles

Jing Han, Binwei Yan, Tianyu Guo et al.

ICML 2025posterarXiv:2512.21708
2
citations

More Than Meets the Eye: Enhancing Multi-Object Tracking Even with Prolonged Occlusions

Bishoy Galoaa, Somaieh Amraee, Sarah Ostadabbas

ICML 2025poster

Morse: Dual-Sampling for Lossless Acceleration of Diffusion Models

Chao Li, Jiawei Fan, Anbang Yao

ICML 2025posterarXiv:2506.18251

MP-Nav: Enhancing Data Poisoning Attacks against Multimodal Learning

Jingfeng Zhang, Prashanth Krishnamurthy, Naman Patel et al.

ICML 2025poster

MPO: An Efficient Post-Processing Framework for Mixing Diverse Preference Alignment

Tianze Wang, Dongnan Gui, Yifan Hu et al.

ICML 2025posterarXiv:2502.18699
4
citations