"natural language processing" Papers

22 papers found

AdaptDel: Adaptable Deletion Rate Randomized Smoothing for Certified Robustness

Zhuoqun Huang, Neil Marchant, Olga Ohrimenko et al.

NEURIPS 2025arXiv:2511.09316

AdaptMI: Adaptive Skill-based In-context Math Instructions for Small Language Models

Yinghui He, Abhishek Panigrahi, Yong Lin et al.

COLM 2025paperarXiv:2505.00147
3
citations

Belief-Calibrated Multi-Agent Consensus Seeking for Complex NLP Tasks

Wentao Deng, Jiahuan Pei, Zhiwei Xu et al.

NEURIPS 2025arXiv:2510.06307

Causal Reasoning and Large Language Models: Opening a New Frontier for Causality

Chenhao Tan, Robert Ness, Amit Sharma et al.

ICLR 2025arXiv:2305.00050
403
citations

Enhancing Transformers Through Conditioned Embedded Tokens

Hemanth Saratchandran, Simon Lucey

ICCV 2025arXiv:2505.12789
2
citations

FormalAlign: Automated Alignment Evaluation for Autoformalization

Jianqiao Lu, Yingjia Wan, Yinya Huang et al.

ICLR 2025arXiv:2410.10135
10
citations

LongProc: Benchmarking Long-Context Language Models on Long Procedural Generation

Xi Ye, Fangcong Yin, Yinghui He et al.

COLM 2025paper
15
citations

MURKA: Multi-Reward Reinforcement Learning with Knowledge Alignment for Optimization Tasks

WANTONG XIE, Yi-Xiang Hu, Jieyang Xu et al.

NEURIPS 2025

Overcoming Long Context Limitations of State Space Models via Context Dependent Sparse Attention

Zhihao Zhan, Jianan Zhao, Zhaocheng Zhu et al.

NEURIPS 2025

Till the Layers Collapse: Compressing a Deep Neural Network Through the Lenses of Batch Normalization Layers.

Zhu Liao, Nour Hezbri, Victor Quétu et al.

AAAI 2025paperarXiv:2412.15077

Torch-Uncertainty: Deep Learning Uncertainty Quantification

Adrien Lafage, Olivier Laurent, Firas Gabetni et al.

NEURIPS 2025spotlight

Zero-Shot Performance Prediction for Probabilistic Scaling Laws

Viktoria Schram, Markus Hiller, Daniel Beck et al.

NEURIPS 2025arXiv:2510.16743

Breaking through the learning plateaus of in-context learning in Transformer

Jingwen Fu, Tao Yang, Yuwang Wang et al.

ICML 2024arXiv:2309.06054
5
citations

Conformal Autoregressive Generation: Beam Search with Coverage Guarantees

Nicolas Deutschmann, Marvin Alberts, María Rodríguez Martínez

AAAI 2024paperarXiv:2309.03797
20
citations

CurBench: Curriculum Learning Benchmark

Yuwei Zhou, Zirui Pan, Xin Wang et al.

ICML 2024

Defense against Backdoor Attack on Pre-trained Language Models via Head Pruning and Attention Normalization

Xingyi Zhao, Depeng Xu, Shuhan Yuan

ICML 2024

Dialogue for Prompting: A Policy-Gradient-Based Discrete Prompt Generation for Few-Shot Learning

Chengzhengxu Li, Xiaoming Liu, Yichen Wang et al.

AAAI 2024paperarXiv:2308.07272
7
citations

Few-Shot Character Understanding in Movies as an Assessment to Meta-Learning of Theory-of-Mind

Mo Yu, Qiujing Wang, Shunchi Zhang et al.

ICML 2024arXiv:2211.04684
21
citations

OptiMUS: Scalable Optimization Modeling with (MI)LP Solvers and Large Language Models

Ali AhmadiTeshnizi, Wenzhi Gao, Madeleine Udell

ICML 2024arXiv:2402.10172
62
citations

Removing Spurious Concepts from Neural Network Representations via Joint Subspace Estimation

Floris Holstege, Bram Wouters, Noud van Giersbergen et al.

ICML 2024arXiv:2310.11991
3
citations

Revisiting Character-level Adversarial Attacks for Language Models

Elias Abad Rocamora, Yongtao Wu, Fanghui Liu et al.

ICML 2024arXiv:2405.04346
6
citations

SpikeZIP-TF: Conversion is All You Need for Transformer-based SNN

kang you, Zekai Xu, Chen Nie et al.

ICML 2024arXiv:2406.03470
20
citations