NEURIPS "large language models" Papers

298 papers found • Page 5 of 6

RAST: Reasoning Activation in LLMs via Small-model Transfer

Siru Ouyang, Xinyu Zhu, Zilin Xiao et al.

NEURIPS 2025posterarXiv:2506.15710
1
citations

Ravan: Multi-Head Low-Rank Adaptation for Federated Fine-Tuning

Arian Raje, Baris Askin, Divyansh Jhunjhunwala et al.

NEURIPS 2025posterarXiv:2506.05568
1
citations

Reasoning Models Better Express Their Confidence

Dongkeun Yoon, Seungone Kim, Sohee Yang et al.

NEURIPS 2025posterarXiv:2505.14489
33
citations

Red-Teaming Text-to-Image Systems by Rule-based Preference Modeling

Yichuan Cao, Yibo Miao, Xiao-Shan Gao et al.

NEURIPS 2025posterarXiv:2505.21074
1
citations

Reinforcement Learning with Backtracking Feedback

Bilgehan Sel, Vaishakh Keshava, Phillip Wallis et al.

NEURIPS 2025poster

Reliable Decision‑Making via Calibration‑Oriented Retrieval‑Augmented Generation

Chaeyun Jang, Deukhwan Cho, Seanie Lee et al.

NEURIPS 2025poster

ReMA: Learning to Meta-Think for LLMs with Multi-agent Reinforcement Learning

Ziyu Wan, Yunxiang Li, Xiaoyu Wen et al.

NEURIPS 2025posterarXiv:2503.09501
36
citations

Reparameterized LLM Training via Orthogonal Equivalence Transformation

Zeju Qiu, Simon Buchholz, Tim Xiao et al.

NEURIPS 2025posterarXiv:2506.08001
2
citations

RepGuard: Adaptive Feature Decoupling for Robust Backdoor Defense in Large Language Models

Chenxu Niu, Jie Zhang, Yanbing Liu et al.

NEURIPS 2025poster

Representation Consistency for Accurate and Coherent LLM Answer Aggregation

Junqi Jiang, Tom Bewley, Salim I. Amoukou et al.

NEURIPS 2025posterarXiv:2506.21590
2
citations

RESAnything: Attribute Prompting for Arbitrary Referring Segmentation

Ruiqi Wang, Hao Zhang

NEURIPS 2025posterarXiv:2505.02867
2
citations

ReSearch: Learning to Reason with Search for LLMs via Reinforcement Learning

Mingyang Chen, Linzhuang Sun, Tianpeng Li et al.

NEURIPS 2025posterarXiv:2503.19470
56
citations

Rethinking Residual Distribution in Locate-then-Edit Model Editing

Xiaopeng Li, Shangwen Wang, Shasha Li et al.

NEURIPS 2025posterarXiv:2502.03748
2
citations

Retro-R1: LLM-based Agentic Retrosynthesis

Wei Liu, Jiangtao Feng, Hongli Yu et al.

NEURIPS 2025poster

Revising and Falsifying Sparse Autoencoder Feature Explanations

George Ma, Samuel Pfrommer, Somayeh Sojoudi

NEURIPS 2025poster

Revolutionizing Training-Free NAS: Towards Efficient Automatic Proxy Discovery via Large Language Models

Haidong Kang, Lihong Lin, Hanling Wang

NEURIPS 2025poster

RHYTHM: Reasoning with Hierarchical Temporal Tokenization for Human Mobility

Haoyu He, Haozheng Luo, Yan Chen et al.

NEURIPS 2025oralarXiv:2509.23115
1
citations

Right Question is Already Half the Answer: Fully Unsupervised LLM Reasoning Incentivization

Qingyang Zhang, Haitao Wu, Changqing Zhang et al.

NEURIPS 2025spotlightarXiv:2504.05812
76
citations

Risk-aware Direct Preference Optimization under Nested Risk Measure

Lijun Zhang, Lin Li, Yajie Qi et al.

NEURIPS 2025posterarXiv:2505.20359
2
citations

Robust Hallucination Detection in LLMs via Adaptive Token Selection

Mengjia Niu, Hamed Haddadi, Guansong Pang

NEURIPS 2025posterarXiv:2504.07863
4
citations

RSAVQ: Riemannian Sensitivity-Aware Vector Quantization for Large Language Models

Zukang Xu, Xing Hu, Qiang Wu et al.

NEURIPS 2025posterarXiv:2510.01240

rStar-Coder: Scaling Competitive Code Reasoning with a Large-Scale Verified Dataset

Yifei Liu, Li Lyna Zhang, Yi Zhu et al.

NEURIPS 2025posterarXiv:2505.21297
22
citations

Scalable Fingerprinting of Large Language Models

Anshul Nasery, Jonathan Hayase, Creston Brooks et al.

NEURIPS 2025spotlightarXiv:2502.07760
8
citations

Scaling and context steer LLMs along the same computational path as the human brain

Joséphine Raugel, Jérémy Rapin, Stéphane d'Ascoli et al.

NEURIPS 2025oralarXiv:2512.01591

scPilot: Large Language Model Reasoning Toward Automated Single-Cell Analysis and Discovery

Yiming Gao, Zhen Wang, Jefferson Chen et al.

NEURIPS 2025poster

Segment Policy Optimization: Effective Segment-Level Credit Assignment in RL for Large Language Models

Yiran Guo, Lijie Xu, Jie Liu et al.

NEURIPS 2025posterarXiv:2505.23564
15
citations

Self-Evolving Pseudo-Rehearsal for Catastrophic Forgetting with Task Similarity in LLMs

Jun Wang, Liang Ding, Shuai Wang et al.

NEURIPS 2025poster

Self Iterative Label Refinement via Robust Unlabeled Learning

Hikaru Asano, Tadashi Kozuno, Yukino Baba

NEURIPS 2025posterarXiv:2502.12565
1
citations

Self-Verification Provably Prevents Model Collapse in Recursive Synthetic Training

Shi Fu, Yingjie Wang, Yuzhu Chen et al.

NEURIPS 2025poster

Semantic-guided Diverse Decoding for Large Language Model

Weijie Shi, Yue Cui, Yaguang Wu et al.

NEURIPS 2025posterarXiv:2506.23601
1
citations

Semantic-KG: Using Knowledge Graphs to Construct Benchmarks for Measuring Semantic Similarity

Qiyao Wei, Edward R Morrell, Lea Goetz et al.

NEURIPS 2025posterarXiv:2511.19925

SeRL: Self-play Reinforcement Learning for Large Language Models with Limited Data

Wenkai Fang, Shunyu Liu, Yang Zhou et al.

NEURIPS 2025posterarXiv:2505.20347
19
citations

ShiQ: Bringing back Bellman to LLMs

Pierre Clavier, Nathan Grinsztajn, Raphaël Avalos et al.

NEURIPS 2025posterarXiv:2505.11081
1
citations

Short-length Adversarial Training Helps LLMs Defend Long-length Jailbreak Attacks: Theoretical and Empirical Evidence

Shaopeng Fu, Liang Ding, Jingfeng ZHANG et al.

NEURIPS 2025posterarXiv:2502.04204
6
citations

SIGMA: Refining Large Language Model Reasoning via Sibling-Guided Monte Carlo Augmentation

Yanwei Ren, Haotian Zhang, Fuxiang Wu et al.

NEURIPS 2025spotlightarXiv:2506.06470

SilentStriker: Toward Stealthy Bit-Flip Attacks on Large Language Models

HAOTIAN XU, Qingsong Peng, Jie Shi et al.

NEURIPS 2025poster
1
citations

Simple and Efficient Heterogeneous Temporal Graph Neural Network

Yili Wang, Tairan Huang, Changlong He et al.

NEURIPS 2025oralarXiv:2510.18467

Simulating Society Requires Simulating Thought

Chance Jiajie Li, Jiayi Wu, Zhenze MO et al.

NEURIPS 2025posterarXiv:2506.06958

Sinusoidal Initialization, Time for a New Start

Alberto Fernandez-Hernandez, Jose Mestre, Manuel F. Dolz et al.

NEURIPS 2025posterarXiv:2505.12909
1
citations

SiriuS: Self-improving Multi-agent Systems via Bootstrapped Reasoning

Wanjia Zhao, Mert Yuksekgonul, Shirley Wu et al.

NEURIPS 2025posterarXiv:2502.04780
22
citations

S'MoRE: Structural Mixture of Residual Experts for Parameter-Efficient LLM Fine-tuning

Hanqing Zeng, Yinglong Xia, Zhuokai Zhao et al.

NEURIPS 2025posterarXiv:2504.06426
2
citations

Solver-Informed RL: Grounding Large Language Models for Authentic Optimization Modeling

Yitian Chen, Jingfan Xia, Siyu Shao et al.

NEURIPS 2025posterarXiv:2505.11792
12
citations

SolverLLM: Leveraging Test-Time Scaling for Optimization Problem via LLM-Guided Search

Dong Li, Xujiang Zhao, Linlin Yu et al.

NEURIPS 2025posterarXiv:2510.16916
1
citations

Sparse MeZO: Less Parameters for Better Performance in Zeroth-Order LLM Fine-Tuning

Yong Liu, Zirui Zhu, Chaoyu Gong et al.

NEURIPS 2025posterarXiv:2402.15751
37
citations

SSTAG: Structure-Aware Self-Supervised Learning Method for Text-Attributed Graphs

Ruyue Liu, Rong Yin, Xiangzhen Bo et al.

NEURIPS 2025posterarXiv:2510.01248
1
citations

SteerConf: Steering LLMs for Confidence Elicitation

Ziang Zhou, Tianyuan Jin, Jieming Shi et al.

NEURIPS 2025posterarXiv:2503.02863
6
citations

Steering When Necessary: Flexible Steering Large Language Models with Backtracking

Zifeng Cheng, Jinwei Gan, Zhiwei Jiang et al.

NEURIPS 2025posterarXiv:2508.17621
1
citations

STEER-ME: Assessing the Microeconomic Reasoning of Large Language Models

Narun Raman, Taylor Lundy, Thiago Amin et al.

NEURIPS 2025posterarXiv:2502.13119
3
citations

Stop DDoS Attacking the Research Community with AI-Generated Survey Papers

Jianghao Lin, Rong Shan, Jiachen Zhu et al.

NEURIPS 2025oralarXiv:2510.09686

Streaming Attention Approximation via Discrepancy Theory

Ekaterina Kochetkova, Kshiteej Jitesh Sheth, Insu Han et al.

NEURIPS 2025spotlightarXiv:2502.07861
2
citations