ICML Papers
5,975 papers found • Page 36 of 120
Looking Beyond the Top-1: Transformers Determine Top Tokens in Order
Daria Lioubashevski, Tomer Schlank, Gabriel Stanovsky et al.
Look Twice Before You Answer: Memory-Space Visual Retracing for Hallucination Mitigation in Multimodal Large Language Models
Xin Zou, Yizhou WANG, Yibo Yan et al.
LoRA-Gen: Specializing Large Language Model via Online LoRA Generation
Yicheng Xiao, Lin Song, Rui Yang et al.
LoRA-One: One-Step Full Gradient Could Suffice for Fine-Tuning Large Language Models, Provably and Efficiently
Yuanhe Zhang, Fanghui Liu, Yudong Chen
LoRA Training Provably Converges to a Low-Rank Global Minimum Or It Fails Loudly (But it Probably Won't Fail)
Junsu Kim, Jaeyeon Kim, Ernest Ryu
Loss Functions and Operators Generated by f-Divergences
Vincent Roulet, Tianlin Liu, Nino Vieillard et al.
LotteryCodec: Searching the Implicit Representation in a Random Network for Low-Complexity Image Compression
Haotian Wu, Gongpu Chen, Pier Luigi Dragotti et al.
Low-Dimension-to-High-Dimension Generalization and Its Implications for Length Generalization
Yang Chen, Long Yang, Yitao Liang et al.
Low-distortion and GPU-compatible Tree Embeddings in Hyperbolic Space
Max van Spengler, Pascal Mettes
Lower Bounds for Chain-of-Thought Reasoning in Hard-Attention Transformers
Alireza Amiribavandpour, Xinting Huang, Mark Rofin et al.
LowRA: Accurate and Efficient LoRA Fine-Tuning of LLMs under 2 Bits
Zikai Zhou, Qizheng Zhang, Hermann Kumbong et al.
Low-Rank Adapting Models for Sparse Autoencoders
Matthew Chen, Josh Engels, Max Tegmark
Low-Rank Tensor Transitions (LoRT) for Transferable Tensor Regression
Andong Wang, Yuning Qiu, Zhong Jin et al.
Low-Rank Thinning
Annabelle Carrell, Albert Gong, Abhishek Shetty et al.
LRA-QViT: Integrating Low-Rank Approximation and Quantization for Robust and Efficient Vision Transformers
Beom Jin Kang, NamJoon Kim, Hyun Kim
LSCD: Lomb--Scargle Conditioned Diffusion for Time series Imputation
Elizabeth M Fons Etcheverry, Alejandro Sztrajman, Yousef El-Laham et al.
LV-XAttn: Distributed Cross-Attention for Long Visual Inputs in Multimodal Large Language Models
Tzu-Tao (Tommy) Chang, Shivaram Venkataraman
M2PDE: Compositional Generative Multiphysics and Multi-component PDE Simulation
Tao Zhang, Zhenhai Liu, Feipeng Qi et al.
M³HF: Multi-agent Reinforcement Learning from Multi-phase Human Feedback of Mixed Quality
Ziyan Wang, Zhicheng Zhang, Fei Fang et al.
M3-JEPA: Multimodal Alignment via Multi-gate MoE based on the Joint-Embedding Predictive Architecture
Hongyang Lei, Xiaolong Cheng, Qi Qin et al.
Machine Learning meets Algebraic Combinatorics: A Suite of Datasets Capturing Research-level Conjecturing Ability in Pure Mathematics
Herman Chau, Helen Jenne, Davis Brown et al.
Machines and Mathematical Mutations: Using GNNs to Characterize Quiver Mutation Classes
Jesse He, Helen Jenne, Herman Chau et al.
MAGELLAN: Metacognitive predictions of learning progress guide autotelic LLM agents in large goal spaces
Loris Gaven, Thomas Carta, Clément Romac et al.
Mahalanobis++: Improving OOD Detection via Feature Normalization
Maximilian Müller, Matthias Hein
Maintaining Proportional Committees with Dynamic Candidate Sets
Chris Dong, Jannik Peters
Make LoRA Great Again: Boosting LoRA with Adaptive Singular Values and Mixture-of-Experts Optimization Alignment
Chenghao Fan, zhenyi lu, Sichen Liu et al.
Making Hard Problems Easier with Custom Data Distributions and Loss Regularization: A Case Study in Modular Arithmetic
Eshika Saxena, Alberto Alfarano, Emily Wenger et al.
MA-LoT: Model-Collaboration Lean-based Long Chain-of-Thought Reasoning enhances Formal Theorem Proving
Ruida Wang, Rui Pan, Yuxin Li et al.
MapEval: A Map-Based Evaluation of Geo-Spatial Reasoning in Foundation Models
Mahir Labib Dihan, Tanvir Hassan, Md Tanvir Parvez et al.
MAPLE: Many-Shot Adaptive Pseudo-Labeling for In-Context Learning
Zihan Chen, Song Wang, Zhen Tan et al.
MARGE: Improving Math Reasoning with Guided Exploration
Jingyue Gao, Runji Lin, Keming Lu et al.
MARS: Unleashing the Power of Variance Reduction for Training Large Models
Huizhuo Yuan, Yifeng Liu, Shuang Wu et al.
MAS-GPT: Training LLMs to Build LLM-based Multi-Agent Systems
Rui Ye, shuo tang, Rui Ge et al.
Masked Autoencoders Are Effective Tokenizers for Diffusion Models
Hao Chen, Yujin Han, Fangyi Chen et al.
Masked Generative Nested Transformers with Decode Time Scaling
Sahil Goyal, Debapriya Tula, Gagan Jain et al.
Mask-Enhanced Autoregressive Prediction: Pay Less Attention to Learn More
Xialie Zhuang, Zhikai Jia, Jianjin Li et al.
MaskTwins: Dual-form Complementary Masking for Domain-Adaptive Image Segmentation
Jiawen Wang, Yinda Chen, Xiaoyu Liu et al.
Massive Values in Self-Attention Modules are the Key to Contextual Knowledge Understanding
Mingyu Jin, Kai Mei, Wujiang Xu et al.
MASS: Mathematical Data Selection via Skill Graphs for Pretraining Large Language Models
Jiazheng Li, Lu Yu, Qing Cui et al.
Mastering Board Games by External and Internal Planning with Language Models
John Schultz, Jakub Adamek, Matej Jusup et al.
Mastering Massive Multi-Task Reinforcement Learning via Mixture-of-Expert Decision Transformer
Yilun Kong, Guozheng Ma, Qi Zhao et al.
Mastering Multiple-Expert Routing: Realizable $H$-Consistency and Strong Guarantees for Learning to Defer
Anqi Mao, Mehryar Mohri, Yutao Zhong
MathConstruct: Challenging LLM Reasoning with Constructive Proofs
Mislav Balunovic, Jasper Dekoninck, Nikola Jovanović et al.
MATH-Perturb: Benchmarking LLMs' Math Reasoning Abilities against Hard Perturbations
Kaixuan Huang, Jiacheng Guo, Zihao Li et al.
Matrix Completion with Incomplete Side Information via Orthogonal Complement Projection
Gengshuo Chang, Wei Zhang, Lehan Zhang
Matryoshka Quantization
Pranav Nair, Puranjay Datta, Jeff Dean et al.
MATS: An Audio Language Model under Text-only Supervision
Wen Wang, Ruibing Hou, Hong Chang et al.
Maximizing Intermediate Checkpoint Value in LLM Pretraining with Bayesian Optimization
Deyuan Liu, Zecheng Wang, Bingning Wang et al.
Maximum Coverage in Turnstile Streams with Applications to Fingerprinting Measures
Alina Ene, Alessandro Epasto, Vahab Mirrokni et al.
Maximum Entropy Reinforcement Learning with Diffusion Policy
Xiaoyi Dong, Jian Cheng, Xi Zhang