Poster "large language models" Papers

740 papers found • Page 1 of 15

$\texttt{G1}$: Teaching LLMs to Reason on Graphs with Reinforcement Learning

Xiaojun Guo, Ang Li, Yifei Wang et al.

NEURIPS 2025
4
citations

3D-AffordanceLLM: Harnessing Large Language Models for Open-Vocabulary Affordance Detection in 3D Worlds

Hengshuo Chu, Xiang Deng, Qi Lv et al.

ICLR 2025arXiv:2502.20041
16
citations

A$^3$E: Towards Compositional Model Editing

Hongming Piao, Hao Wang, Dapeng Wu et al.

NEURIPS 2025

ACC-Collab: An Actor-Critic Approach to Multi-Agent LLM Collaboration

Andrew Estornell, Jean-Francois Ton, Yuanshun Yao et al.

ICLR 2025arXiv:2411.00053
13
citations

Accelerating Block Coordinate Descent for LLM Finetuning via Landscape Expansion

Qijun Luo, Yifei Shen, Liangzu Peng et al.

NEURIPS 2025

Accelerating RL for LLM Reasoning with Optimal Advantage Regression

Kianté Brantley, Mingyu Chen, Zhaolin Gao et al.

NEURIPS 2025arXiv:2505.20686
12
citations

A Closer Look at Machine Unlearning for Large Language Models

Xiaojian Yuan, Tianyu Pang, Chao Du et al.

ICLR 2025arXiv:2410.08109
35
citations

ActionReasoningBench: Reasoning about Actions with and without Ramification Constraints

Divij Handa, Pavel Dolin, Shrinidhi Kumbhar et al.

ICLR 2025arXiv:2406.04046
7
citations

Activation-Guided Consensus Merging for Large Language Models

Yuxuan Yao, Shuqi LIU, Zehua Liu et al.

NEURIPS 2025arXiv:2505.14009
2
citations

AcuRank: Uncertainty-Aware Adaptive Computation for Listwise Reranking

Soyoung Yoon, Gyuwan Kim, Gyu-Hwung Cho et al.

NEURIPS 2025arXiv:2505.18512
2
citations

Ada-K Routing: Boosting the Efficiency of MoE-based LLMs

Zijia Zhao, Longteng Guo, Jie Cheng et al.

ICLR 2025arXiv:2410.10456
8
citations

AdaLRS: Loss-Guided Adaptive Learning Rate Search for Efficient Foundation Model Pretraining

Hongyuan Dong, Dingkang Yang, Xiao Liang et al.

NEURIPS 2025arXiv:2506.13274
3
citations

Adaptive Distraction: Probing LLM Contextual Robustness with Automated Tree Search

Yanbo Wang, Zixiang Xu, Yue Huang et al.

NEURIPS 2025arXiv:2502.01609
3
citations

Adaptive Kernel Design for Bayesian Optimization Is a Piece of CAKE with LLMs

Richard Suwandi, Feng Yin, Juntao Wang et al.

NEURIPS 2025arXiv:2509.17998
2
citations

Addax: Utilizing Zeroth-Order Gradients to Improve Memory Efficiency and Performance of SGD for Fine-Tuning Language Models

Zeman Li, Xinwei Zhang, Peilin Zhong et al.

ICLR 2025arXiv:2410.06441
11
citations

AdmTree: Compressing Lengthy Context with Adaptive Semantic Trees

Yangning Li, Shaoshen Chen, Yinghui Li et al.

NEURIPS 2025arXiv:2512.04550
4
citations

Advancing LLM Reasoning Generalists with Preference Trees

Lifan Yuan, Ganqu Cui, Hanbin Wang et al.

ICLR 2025arXiv:2404.02078
183
citations

AdvPrompter: Fast Adaptive Adversarial Prompting for LLMs

Anselm Paulus, Arman Zharmagambetov, Chuan Guo et al.

ICML 2025arXiv:2404.16873
132
citations

Afterburner: Reinforcement Learning Facilitates Self-Improving Code Efficiency Optimization

Mingzhe Du, Anh Tuan Luu, Yue Liu et al.

NEURIPS 2025arXiv:2505.23387
7
citations

Agents' Room: Narrative Generation through Multi-step Collaboration

Fantine Huot, Reinald Kim Amplayo, Jennimaria Palomaki et al.

ICLR 2025arXiv:2410.02603
42
citations

AgentTTS: Large Language Model Agent for Test-time Compute-optimal Scaling Strategy in Complex Tasks

Fali Wang, Hui Liu, Zhenwei Dai et al.

NEURIPS 2025arXiv:2508.00890
9
citations

AI as Humanity’s Salieri: Quantifying Linguistic Creativity of Language Models via Systematic Attribution of Machine Text against Web Text

Ximing Lu, Melanie Sclar, Skyler Hallinan et al.

ICLR 2025arXiv:2410.04265
33
citations

AIMS.au: A Dataset for the Analysis of Modern Slavery Countermeasures in Corporate Statements

Adriana-Eufrosina Bora, Pierre-Luc St-Charles, Mirko Bronzi et al.

ICLR 2025arXiv:2502.07022
2
citations

Alignment of Large Language Models with Constrained Learning

Botong Zhang, Shuo Li, Ignacio Hounie et al.

NEURIPS 2025arXiv:2505.19387
2
citations

ALLaM: Large Language Models for Arabic and English

M Saiful Bari, Yazeed Alnumay, Norah Alzahrani et al.

ICLR 2025arXiv:2407.15390
49
citations

Alleviating Hallucinations in Large Language Models through Multi-Model Contrastive Decoding and Dynamic Hallucination Detection

Chenyu Zhu, Yefeng Liu, Hao Zhang et al.

NEURIPS 2025

AlphaDecay: Module-wise Weight Decay for Heavy-Tailed Balancing in LLMs

Di He, Songjun Tu, Ajay Jaiswal et al.

NEURIPS 2025arXiv:2506.14562
1
citations

A Multi-Power Law for Loss Curve Prediction Across Learning Rate Schedules

Kairong Luo, Haodong Wen, Shengding Hu et al.

ICLR 2025arXiv:2503.12811
17
citations

Analyzing the Power of Chain of Thought through Memorization Capabilities

Lijia Yu, Xiao-Shan Gao, Lijun Zhang

NEURIPS 2025arXiv:2511.01190

An Intelligent Agentic System for Complex Image Restoration Problems

Kaiwen Zhu, Jinjin Gu, Zhiyuan You et al.

ICLR 2025arXiv:2410.17809
26
citations

AnoLLM: Large Language Models for Tabular Anomaly Detection

Che-Ping Tsai, Ganyu Teng, Phillip Wallis et al.

ICLR 2025
7
citations

API Pack: A Massive Multi-Programming Language Dataset for API Call Generation

Gavin (Zhen) Guo, Adriana Meza Soria, Wei Sun et al.

ICLR 2025arXiv:2402.09615
4
citations

Approximately Aligned Decoding

Daniel Melcer, Sujan Kumar Gonugondla, Pramuditha Perera et al.

NEURIPS 2025arXiv:2410.01103
2
citations

A Probabilistic Perspective on Unlearning and Alignment for Large Language Models

Yan Scholten, Stephan Günnemann, Leo Schwinn

ICLR 2025arXiv:2410.03523
17
citations

AREAL: A Large-Scale Asynchronous Reinforcement Learning System for Language Reasoning

Wei Fu, Jiaxuan Gao, Xujie Shen et al.

NEURIPS 2025arXiv:2505.24298
117
citations

Are Sparse Autoencoders Useful? A Case Study in Sparse Probing

Subhash Kantamneni, Josh Engels, Senthooran Rajamanoharan et al.

ICML 2025arXiv:2502.16681
56
citations

A Simple yet Effective Layout Token in Large Language Models for Document Understanding

Zhaoqing Zhu, Chuwei Luo, Zirui Shao et al.

CVPR 2025arXiv:2503.18434
7
citations

Ask, and it shall be given: On the Turing completeness of prompting

Ruizhong Qiu, Zhe Xu, Wenxuan Bao et al.

ICLR 2025arXiv:2411.01992
6
citations

A Statistical Approach for Controlled Training Data Detection

Zirui Hu, Yingjie Wang, Zheng Zhang et al.

ICLR 2025
2
citations

ATLAS: Autoformalizing Theorems through Lifting, Augmentation, and Synthesis of Data

Xiaoyang Liu, Kangjie Bao, Jiashuo Zhang et al.

NEURIPS 2025arXiv:2502.05567
13
citations

A Training-Free Sub-quadratic Cost Transformer Model Serving Framework with Hierarchically Pruned Attention

Heejun Lee, Geon Park, Youngwan Lee et al.

ICLR 2025arXiv:2406.09827
9
citations

AttriBoT: A Bag of Tricks for Efficiently Approximating Leave-One-Out Context Attribution

Fengyuan Liu, Nikhil Kandpal, Colin Raffel

ICLR 2025arXiv:2411.15102
15
citations

AutoData: A Multi-Agent System for Open Web Data Collection

Tianyi Ma, Yiyue Qian, Zheyuan Zhang et al.

NEURIPS 2025arXiv:2505.15859
5
citations

Automatic Auxiliary Task Selection and Adaptive Weighting Boost Molecular Property Prediction

Zhiqiang Zhong, Davide Mottin

NEURIPS 2025

Autonomous Evaluation of LLMs for Truth Maintenance and Reasoning Tasks

Rushang Karia, Daniel Bramblett, Daksh Dobhal et al.

ICLR 2025arXiv:2410.08437
2
citations

AutoPrompt: Automated Red-Teaming of Text-to-Image Models via LLM-Driven Adversarial Prompts

Yufan Liu, Wanqian Zhang, Huashan Chen et al.

ICCV 2025arXiv:2510.24034
1
citations

AutoRedTeamer: Autonomous Red Teaming with Lifelong Attack Integration

Andy Zhou, Kevin Wu, Francesco Pinto et al.

NEURIPS 2025arXiv:2503.15754
16
citations

Basis Sharing: Cross-Layer Parameter Sharing for Large Language Model Compression

Jingcun Wang, Yu-Guang Chen, Ing-Chao Lin et al.

ICLR 2025arXiv:2410.03765
21
citations

Bayesian Concept Bottleneck Models with LLM Priors

Jean Feng, Avni Kothari, Lucas Zier et al.

NEURIPS 2025arXiv:2410.15555
10
citations

Better autoregressive regression with LLMs via regression-aware fine-tuning

Michal Lukasik, Zhao Meng, Harikrishna Narasimhan et al.

ICLR 2025
7
citations
PreviousNext