NEURIPS "large language models" Papers
298 papers found • Page 1 of 6
$\texttt{G1}$: Teaching LLMs to Reason on Graphs with Reinforcement Learning
Xiaojun Guo, Ang Li, Yifei Wang et al.
A$^3$E: Towards Compositional Model Editing
Hongming Piao, Hao Wang, Dapeng Wu et al.
Accelerating Block Coordinate Descent for LLM Finetuning via Landscape Expansion
Qijun Luo, Yifei Shen, Liangzu Peng et al.
Accelerating RL for LLM Reasoning with Optimal Advantage Regression
Kianté Brantley, Mingyu Chen, Zhaolin Gao et al.
Activation-Guided Consensus Merging for Large Language Models
Yuxuan Yao, Shuqi LIU, Zehua Liu et al.
AcuRank: Uncertainty-Aware Adaptive Computation for Listwise Reranking
Soyoung Yoon, Gyuwan Kim, Gyu-Hwung Cho et al.
AdaLRS: Loss-Guided Adaptive Learning Rate Search for Efficient Foundation Model Pretraining
Hongyuan Dong, Dingkang Yang, Xiao Liang et al.
Adaptive Distraction: Probing LLM Contextual Robustness with Automated Tree Search
Yanbo Wang, Zixiang Xu, Yue Huang et al.
Adaptive Kernel Design for Bayesian Optimization Is a Piece of CAKE with LLMs
Richard Suwandi, Feng Yin, Juntao Wang et al.
Adaptive Prediction-Powered AutoEval with Reliability and Efficiency Guarantees
Sangwoo Park, Matteo Zecchin, Osvaldo Simeone
AdmTree: Compressing Lengthy Context with Adaptive Semantic Trees
Yangning Li, Shaoshen Chen, Yinghui Li et al.
Advancing Expert Specialization for Better MoE
Hongcan Guo, Haolang Lu, Guoshun Nan et al.
Afterburner: Reinforcement Learning Facilitates Self-Improving Code Efficiency Optimization
Mingzhe Du, Anh Tuan Luu, Yue Liu et al.
AGENTIF: Benchmarking Large Language Models Instruction Following Ability in Agentic Scenarios
Yunjia Qi, Hao Peng, Xiaozhi Wang et al.
AgentTTS: Large Language Model Agent for Test-time Compute-optimal Scaling Strategy in Complex Tasks
Fali Wang, Hui Liu, Zhenwei Dai et al.
A Implies B: Circuit Analysis in LLMs for Propositional Logical Reasoning
Guan Zhe Hong, Nishanth Dikkala, Enming Luo et al.
AI-Researcher: Autonomous Scientific Innovation
Jiabin Tang, Lianghao Xia, Zhonghang Li et al.
Alignment of Large Language Models with Constrained Learning
Botong Zhang, Shuo Li, Ignacio Hounie et al.
Alleviating Hallucinations in Large Language Models through Multi-Model Contrastive Decoding and Dynamic Hallucination Detection
Chenyu Zhu, Yefeng Liu, Hao Zhang et al.
AlphaDecay: Module-wise Weight Decay for Heavy-Tailed Balancing in LLMs
Di He, Songjun Tu, Ajay Jaiswal et al.
Analyzing the Power of Chain of Thought through Memorization Capabilities
Lijia Yu, Xiao-Shan Gao, Lijun Zhang
Angles Don’t Lie: Unlocking Training‑Efficient RL Through the Model’s Own Signals
Qinsi Wang, Jinghan Ke, Hancheng Ye et al.
Approximately Aligned Decoding
Daniel Melcer, Sujan Kumar Gonugondla, Pramuditha Perera et al.
AREAL: A Large-Scale Asynchronous Reinforcement Learning System for Language Reasoning
Wei Fu, Jiaxuan Gao, Xujie Shen et al.
ATLAS: Autoformalizing Theorems through Lifting, Augmentation, and Synthesis of Data
Xiaoyang Liu, Kangjie Bao, Jiashuo Zhang et al.
AutoData: A Multi-Agent System for Open Web Data Collection
Tianyi Ma, Yiyue Qian, Zheyuan Zhang et al.
AutoDiscovery: Open-ended Scientific Discovery via Bayesian Surprise
Dhruv Agarwal, Bodhisattwa Prasad Majumder, Reece Adamson et al.
Automatic Auxiliary Task Selection and Adaptive Weighting Boost Molecular Property Prediction
Zhiqiang Zhong, Davide Mottin
AutoRedTeamer: Autonomous Red Teaming with Lifelong Attack Integration
Andy Zhou, Kevin Wu, Francesco Pinto et al.
AutoToM: Scaling Model-based Mental Inference via Automated Agent Modeling
Zhining Zhang, Chuanyang Jin, Mung Yao Jia et al.
Bayesian Concept Bottleneck Models with LLM Priors
Jean Feng, Avni Kothari, Lucas Zier et al.
Beyond Single-Task: Robust Multi-Task Length Generalization for LLMs
Yi Hu, Shijia Kang, Haotong Yang et al.
Bilevel ZOFO: Efficient LLM Fine-Tuning and Meta-Training
Reza Shirkavand, Peiran Yu, Qi He et al.
Bits Leaked per Query: Information-Theoretic Bounds for Adversarial Attacks on LLMs
Masahiro Kaneko, Timothy Baldwin
Boosting Skeleton-based Zero-Shot Action Recognition with Training-Free Test-Time Adaptation
Jingmin Zhu, Anqi Zhu, Hossein Rahmani et al.
Breaking the Gradient Barrier: Unveiling Large Language Models for Strategic Classification
Xinpeng Lv, Yunxin Mao, Haoxuan Li et al.
Bridging Sign and Spoken Languages: Pseudo Gloss Generation for Sign Language Translation
Jianyuan Guo, Peike Li, Trevor Cohn
Bridging Time and Linguistics: LLMs as Time Series Analyzer through Symbolization and Segmentation
Jianyang Qin, Chaoyang Li, Jinhao Cui et al.
C3PO: Optimized Large Language Model Cascades with Probabilistic Cost Constraints for Reasoning
Antonios Valkanas, Soumyasundar Pal, Pavel Rumiantsev et al.
Calibrating Translation Decoding with Quality Estimation on LLMs
Di Wu, Yibin Lei, Christof Monz
Can DPO Learn Diverse Human Values? A Theoretical Scaling Law
Shawn Im, Sharon Li
Can Large Language Models Help Multimodal Language Analysis? MMLA: A Comprehensive Benchmark
Hanlei Zhang, zhuohang li, Hua Xu et al.
Can LLMs Outshine Conventional Recommenders? A Comparative Evaluation
Qijiong Liu, Jieming Zhu, Lu Fan et al.
Can LLMs Reason Over Non-Text Modalities in a Training-Free Manner? A Case Study with In-Context Representation Learning
Tianle Zhang, Wanlong Fang, Jonathan Woo et al.
Can We Infer Confidential Properties of Training Data from LLMs?
Pengrun Huang, Chhavi Yadav, Kamalika Chaudhuri et al.
Capturing Polysemanticity with PRISM: A Multi-Concept Feature Description Framework
Laura Kopf, Nils Feldhus, Kirill Bykov et al.
CCL: Causal-aware In-context Learning for Out-of-Distribution Generalization
Hoyoon Byun, Gyeongdeok Seo, Joonseong Kang et al.
CellVerse: Do Large Language Models Really Understand Cell Biology?
Fan Zhang, Tianyu Liu, Zhihong Zhu et al.
Chain of Execution Supervision Promotes General Reasoning in Large Language Models
Nuo Chen, Zehua Li, Keqin Bao et al.
CIDD: Collaborative Intelligence for Structure-Based Drug Design Empowered by LLMs
Bowen Gao, Yanwen Huang, Yiqiao Liu et al.