ICML "large language models" Papers

180 papers found • Page 4 of 4

Soft Prompt Recovers Compressed LLMs, Transferably

Zhaozhuo Xu, Zirui Liu, Beidi Chen et al.

ICML 2024poster

SPP: Sparsity-Preserved Parameter-Efficient Fine-Tuning for Large Language Models

Xudong LU, Aojun Zhou, Yuhui Xu et al.

ICML 2024posterarXiv:2405.16057

SqueezeLLM: Dense-and-Sparse Quantization

Sehoon Kim, Coleman Hooper, Amir Gholaminejad et al.

ICML 2024posterarXiv:2306.07629

StackSight: Unveiling WebAssembly through Large Language Models and Neurosymbolic Chain-of-Thought Decompilation

Weike Fang, Zhejian Zhou, Junzhou He et al.

ICML 2024spotlightarXiv:2406.04568

Structured Chemistry Reasoning with Large Language Models

Siru Ouyang, Zhuosheng Zhang, Bing Yan et al.

ICML 2024posterarXiv:2311.09656

Subgoal-based Demonstration Learning for Formal Theorem Proving

Xueliang Zhao, Wenda Li, Lingpeng Kong

ICML 2024posterarXiv:2305.16366

Tandem Transformers for Inference Efficient LLMs

Aishwarya P S, Pranav Nair, Yashas Samaga et al.

ICML 2024posterarXiv:2402.08644

The Illusion of State in State-Space Models

William Merrill, Jackson Petty, Ashish Sabharwal

ICML 2024posterarXiv:2404.08819

Thermometer: Towards Universal Calibration for Large Language Models

Maohao Shen, Subhro Das, Kristjan Greenewald et al.

ICML 2024posterarXiv:2403.08819

tinyBenchmarks: evaluating LLMs with fewer examples

Felipe Maia Polo, Lucas Weber, Leshem Choshen et al.

ICML 2024posterarXiv:2402.14992

tnGPS: Discovering Unknown Tensor Network Structure Search Algorithms via Large Language Models (LLMs)

Junhua Zeng, Chao Li, Zhun Sun et al.

ICML 2024posterarXiv:2402.02456

To Cool or not to Cool? Temperature Network Meets Large Foundation Models via DRO

Zi-Hao Qiu, Siqi Guo, Mao Xu et al.

ICML 2024posterarXiv:2404.04575

To Each (Textual Sequence) Its Own: Improving Memorized-Data Unlearning in Large Language Models

George-Octavian Bărbulescu, Peter Triantafillou

ICML 2024poster

Token-level Direct Preference Optimization

Yongcheng Zeng, Guoqing Liu, Weiyu Ma et al.

ICML 2024posterarXiv:2404.11999

Token-Specific Watermarking with Enhanced Detectability and Semantic Coherence for Large Language Models

Mingjia Huo, Sai Ashish Somayajula, Youwei Liang et al.

ICML 2024posterarXiv:2402.18059

Toward Adaptive Reasoning in Large Language Models with Thought Rollback

Sijia Chen, Baochun Li

ICML 2024posterarXiv:2412.19707

Towards AutoAI: Optimizing a Machine Learning System with Black-box and Differentiable Components

Zhiliang Chen, Chuan-Sheng Foo, Bryan Kian Hsiang Low

ICML 2024poster

Trainable Transformer in Transformer

Abhishek Panigrahi, Sadhika Malladi, Mengzhou Xia et al.

ICML 2024posterarXiv:2307.01189

Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning

Zhiheng Xi, Wenxiang Chen, Boyang Hong et al.

ICML 2024posterarXiv:2402.05808

TravelPlanner: A Benchmark for Real-World Planning with Language Agents

Jian Xie, Kai Zhang, Jiangjie Chen et al.

ICML 2024spotlightarXiv:2402.01622

ULTRAFEEDBACK: Boosting Language Models with Scaled AI Feedback

Ganqu Cui, Lifan Yuan, Ning Ding et al.

ICML 2024posterarXiv:2310.01377

Understanding the Learning Dynamics of Alignment with Human Feedback

Shawn Im, Sharon Li

ICML 2024posterarXiv:2403.18742

UniAudio: Towards Universal Audio Generation with Large Language Models

Dongchao Yang, Jinchuan Tian, Xu Tan et al.

ICML 2024poster

Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration

Zhongzhi Yu, Zheng Wang, Yonggan Fu et al.

ICML 2024posterarXiv:2406.15765

Unveiling the Potential of AI for Nanomaterial Morphology Prediction

Ivan Dubrovsky, Andrei Dmitrenko, Aleksey Dmitrenko et al.

ICML 2024posterarXiv:2406.02591

Use Your INSTINCT: INSTruction optimization for LLMs usIng Neural bandits Coupled with Transformers

Xiaoqiang Lin, Zhaoxuan Wu, Zhongxiang Dai et al.

ICML 2024posterarXiv:2310.02905

Variational Learning is Effective for Large Deep Networks

Yuesong Shen, Nico Daheim, Bai Cong et al.

ICML 2024spotlightarXiv:2402.17641

Watermark Stealing in Large Language Models

Nikola Jovanović, Robin Staab, Martin Vechev

ICML 2024posterarXiv:2402.19361

When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models

Haoran You, Yichao Fu, Zheng Wang et al.

ICML 2024posterarXiv:2406.07368

Zero-Shot ECG Classification with Multimodal Learning and Test-time Clinical Knowledge Enhancement

che liu, Zhongwei Wan, Cheng Ouyang et al.

ICML 2024posterarXiv:2403.06659