NeurIPS 2025 "computational efficiency" Papers

20 papers found

Accelerating Multimodal Large Language Models via Dynamic Visual-Token Exit and the Empirical Findings

Qiong Wu, Wenhao Lin, Yiyi Zhou et al.

NeurIPS 2025posterarXiv:2411.19628
5
citations

Accurate and Efficient Low-Rank Model Merging in Core Space

Aniello Panariello, Daniel Marczak, Simone Magistri et al.

NeurIPS 2025posterarXiv:2509.17786
3
citations

Adaptive Inference-Time Scaling via Cyclic Diffusion Search

Gyubin Lee, Bao Truong, Jaesik Yoon et al.

NeurIPS 2025posterarXiv:2505.14036

Balanced Token Pruning: Accelerating Vision Language Models Beyond Local Optimization

kaiyuan Li, Xiaoyue Chen, Chen Gao et al.

NeurIPS 2025posterarXiv:2505.22038
4
citations

Beyond Greedy Exits: Improved Early Exit Decisions for Risk Control and Reliability

Divya Jyoti Bajpai, Manjesh Kumar Hanawal

NeurIPS 2025posterarXiv:2509.23666

Bio-Inspired Image Restoration

Yuning Cui, Wenqi Ren, Alois Knoll

NeurIPS 2025poster

Efficient RAW Image Deblurring with Adaptive Frequency Modulation

Wenlong Jiao, Binglong Li, Wei Shang et al.

NeurIPS 2025posterarXiv:2505.24407
1
citations

Faithful Group Shapley Value

Kiljae Lee, Ziqi Liu, Weijing Tang et al.

NeurIPS 2025posterarXiv:2505.19013

Fast Solvers for Discrete Diffusion Models: Theory and Applications of High-Order Algorithms

Yinuo Ren, Haoxuan Chen, Yuchen Zhu et al.

NeurIPS 2025posterarXiv:2502.00234
29
citations

FastVID: Dynamic Density Pruning for Fast Video Large Language Models

Leqi Shen, Guoqiang Gong, Tao He et al.

NeurIPS 2025oralarXiv:2503.11187
16
citations

Gatekeeper: Improving Model Cascades Through Confidence Tuning

Stephan Rabanser, Nathalie Rauschmayr, Achin Kulshrestha et al.

NeurIPS 2025posterarXiv:2502.19335
4
citations

HoliTom: Holistic Token Merging for Fast Video Large Language Models

Kele Shao, Keda TAO, Can Qin et al.

NeurIPS 2025oralarXiv:2505.21334
18
citations

Multi-Agent Collaboration via Evolving Orchestration

Yufan Dang, Chen Qian, Xueheng Luo et al.

NeurIPS 2025posterarXiv:2505.19591
25
citations

One Head to Rule Them All: Amplifying LVLM Safety through a Single Critical Attention Head

Junhao Xia, Haotian Zhu, Shuchao Pang et al.

NeurIPS 2025poster

Robust Regression of General ReLUs with Queries

Ilias Diakonikolas, Daniel Kane, Mingchen Ma

NeurIPS 2025poster

SCOPE: Saliency-Coverage Oriented Token Pruning for Efficient Multimodel LLMs

Jinhong Deng, Wen Li, Joey Tianyi Zhou et al.

NeurIPS 2025posterarXiv:2510.24214

SHF: Symmetrical Hierarchical Forest with Pretrained Vision Transformer Encoder for High-Resolution Medical Segmentation

Enzhi Zhang, Peng Chen, Rui Zhong et al.

NeurIPS 2025spotlight

VCM: Vision Concept Modeling with Adaptive Vision Token Compression via Instruction Fine-Tuning

Run Luo, Renke Shan, Longze Chen et al.

NeurIPS 2025poster

Vision-centric Token Compression in Large Language Model

Ling Xing, Alex Jinpeng Wang, Rui Yan et al.

NeurIPS 2025spotlightarXiv:2502.00791
7
citations

VORTA: Efficient Video Diffusion via Routing Sparse Attention

Wenhao Sun, Rong-Cheng Tu, Yifu Ding et al.

NeurIPS 2025posterarXiv:2505.18809
7
citations