NeurIPS "large language models" Papers

231 papers found • Page 5 of 5

The Best Instruction-Tuning Data are Those That Fit

Dylan Zhang, Qirun Dai, Hao Peng

NeurIPS 2025spotlightarXiv:2502.04194
22
citations

The Common Pile v0.1: An 8TB Dataset of Public Domain and Openly Licensed Text

Nikhil Kandpal, Brian Lester, Colin Raffel et al.

NeurIPS 2025posterarXiv:2506.05209
10
citations

The Rise of Parameter Specialization for Knowledge Storage in Large Language Models

Yihuai Hong, Yiran Zhao, Wei Tang et al.

NeurIPS 2025posterarXiv:2505.17260
1
citations

ThinkBench: Dynamic Out-of-Distribution Evaluation for Robust LLM Reasoning

Shulin Huang, Linyi Yang, Yan Song et al.

NeurIPS 2025posterarXiv:2502.16268
14
citations

Thinker: Learning to Think Fast and Slow

Stephen Chung, Wenyu Du, Jie Fu

NeurIPS 2025posterarXiv:2505.21097
5
citations

Timely Clinical Diagnosis through Active Test Selection

Silas Ruhrberg Estévez, Nicolás Astorga, Mihaela van der Schaar

NeurIPS 2025posterarXiv:2510.18988

Token-Level Self-Play with Importance-Aware Guidance for Large Language Models

Tue Le, Hoang Tran, Quyen Tran et al.

NeurIPS 2025poster

Towards Understanding Safety Alignment: A Mechanistic Perspective from Safety Neurons

Jianhui Chen, Xiaozhi Wang, Zijun Yao et al.

NeurIPS 2025posterarXiv:2406.14144
24
citations

Training-Free Bayesianization for Low-Rank Adapters of Large Language Models

Haizhou Shi, Yibin Wang, Ligong Han et al.

NeurIPS 2025posterarXiv:2412.05723
2
citations

TrajAgent: An LLM-Agent Framework for Trajectory Modeling via Large-and-Small Model Collaboration

Yuwei Du, Jie Feng, Jie Zhao et al.

NeurIPS 2025posterarXiv:2410.20445
3
citations

Tree of Preferences for Diversified Recommendation

Hanyang Yuan, Ning Tang, Tongya Zheng et al.

NeurIPS 2025posterarXiv:2601.02386

Triplets Better Than Pairs: Towards Stable and Effective Self-Play Fine-Tuning for LLMs

Yibo Wang, Hai-Long Sun, Guangda Huzhang et al.

NeurIPS 2025posterarXiv:2601.08198
4
citations

Truth over Tricks: Measuring and Mitigating Shortcut Learning in Misinformation Detection

Herun Wan, Jiaying Wu, Minnan Luo et al.

NeurIPS 2025posterarXiv:2506.02350
6
citations

TSENOR: Highly-Efficient Algorithm for Finding Transposable N:M Sparse Masks

Xiang Meng, Mehdi Makni, Rahul Mazumder

NeurIPS 2025poster

TTRL: Test-Time Reinforcement Learning

Yuxin Zuo, Kaiyan Zhang, Li Sheng et al.

NeurIPS 2025posterarXiv:2504.16084
122
citations

Týr-the-Pruner: Structural Pruning LLMs via Global Sparsity Distribution Optimization

Guanchen Li, Yixing Xu, Zeping Li et al.

NeurIPS 2025posterarXiv:2503.09657
6
citations

Understanding Parametric and Contextual Knowledge Reconciliation within Large Language Models

Jun Zhao, Yongzhuo Yang, Xiang Hu et al.

NeurIPS 2025spotlight

Unifying Text Semantics and Graph Structures for Temporal Text-attributed Graphs with Large Language Models

Siwei Zhang, Yun Xiong, Yateng Tang et al.

NeurIPS 2025oralarXiv:2503.14411
2
citations

Uni-LoRA: One Vector is All You Need

Kaiyang Li, Shaobo Han, Qing Su et al.

NeurIPS 2025spotlightarXiv:2506.00799
2
citations

Unlearned but Not Forgotten: Data Extraction after Exact Unlearning in LLM

Xiaoyu Wu, Yifei Pang, Terrance Liu et al.

NeurIPS 2025posterarXiv:2505.24379
2
citations

VADTree: Explainable Training-Free Video Anomaly Detection via Hierarchical Granularity-Aware Tree

Wenlong Li, Yifei Xu, Yuan Rao et al.

NeurIPS 2025oralarXiv:2510.22693
1
citations

Valid Inference with Imperfect Synthetic Data

Yewon Byun, Shantanu Gupta, Zachary Lipton et al.

NeurIPS 2025posterarXiv:2508.06635

Variational Uncertainty Decomposition for In-Context Learning

I. Shavindra Jayasekera, Jacob Si, Filippo Valdettaro et al.

NeurIPS 2025posterarXiv:2509.02327
1
citations

VERA: Variational Inference Framework for Jailbreaking Large Language Models

Anamika Lochab, Lu Yan, Patrick Pynadath et al.

NeurIPS 2025posterarXiv:2506.22666

Virus Infection Attack on LLMs: Your Poisoning Can Spread "VIA" Synthetic Data

Zi Liang, Qingqing Ye, Xuan Liu et al.

NeurIPS 2025spotlight

VT-FSL: Bridging Vision and Text with LLMs for Few-Shot Learning

Wenhao Li, Qiangchang Wang, Xianjing Meng et al.

NeurIPS 2025posterarXiv:2509.25033
2
citations

What Happens During the Loss Plateau? Understanding Abrupt Learning in Transformers

Pulkit Gopalani, Wei Hu

NeurIPS 2025posterarXiv:2506.13688
1
citations

When Can Model-Free Reinforcement Learning be Enough for Thinking?

Josiah Hanna, Nicholas Corrado

NeurIPS 2025posterarXiv:2506.17124

Why Knowledge Distillation Works in Generative Models: A Minimal Working Explanation

Sungmin Cha, Kyunghyun Cho

NeurIPS 2025posterarXiv:2505.13111
4
citations

Wider or Deeper? Scaling LLM Inference-Time Compute with Adaptive Branching Tree Search

Yuichi Inoue, Kou Misaki, Yuki Imajuku et al.

NeurIPS 2025spotlightarXiv:2503.04412
18
citations

WritingBench: A Comprehensive Benchmark for Generative Writing

Yuning Wu, Jiahao Mei, Ming Yan et al.

NeurIPS 2025posterarXiv:2503.05244
41
citations
← Previous
1...345
Next →