ICML "reinforcement learning" Papers

74 papers found • Page 2 of 2

Reinforcement Learning within Tree Search for Fast Macro Placement

Zijie Geng, Jie Wang, Ziyan Liu et al.

ICML 2024poster

Remembering to Be Fair: Non-Markovian Fairness in Sequential Decision Making

Parand A. Alamdari, Toryn Q. Klassen, Elliot Creager et al.

ICML 2024posterarXiv:2312.04772

Rethinking Transformers in Solving POMDPs

Chenhao Lu, Ruizhe Shi, Yuyao Liu et al.

ICML 2024posterarXiv:2405.17358

Revisiting Scalable Hessian Diagonal Approximations for Applications in Reinforcement Learning

Mohamed Elsayed, Homayoon Farrahi, Felix Dangel et al.

ICML 2024posterarXiv:2406.03276

Reward Shaping for Reinforcement Learning with An Assistant Reward Agent

Haozhe Ma, Kuankuan Sima, Thanh Vinh Vo et al.

ICML 2024poster

RICE: Breaking Through the Training Bottlenecks of Reinforcement Learning with Explanation

Zelei Cheng, Xian Wu, Jiahao Yu et al.

ICML 2024spotlightarXiv:2405.03064

Rich-Observation Reinforcement Learning with Continuous Latent Dynamics

Yuda Song, Lili Wu, Dylan Foster et al.

ICML 2024posterarXiv:2405.19269

Risk-Sensitive Policy Optimization via Predictive CVaR Policy Gradient

Ju-Hyun Kim, Seungki Min

ICML 2024poster

RL-CFR: Improving Action Abstraction for Imperfect Information Extensive-Form Games with Reinforcement Learning

Boning Li, Zhixuan Fang, Longbo Huang

ICML 2024posterarXiv:2403.04344

RoboGen: Towards Unleashing Infinite Data for Automated Robot Learning via Generative Simulation

Yufei Wang, Zhou Xian, Feng Chen et al.

ICML 2024posterarXiv:2311.01455

Robust Optimization in Protein Fitness Landscapes Using Reinforcement Learning in Latent Space

Minji Lee, Luiz Felipe Vecchietti, Hyunkyu Jung et al.

ICML 2024spotlightarXiv:2405.18986

Run-Time Task Composition with Safety Semantics

Kevin Leahy, Makai Mann, Zachary Serlin

ICML 2024poster

Sample Average Approximation for Conditional Stochastic Optimization with Dependent Data

Yafei Wang, Bo Pan, Mei Li et al.

ICML 2024poster

SiT: Symmetry-invariant Transformers for Generalisation in Reinforcement Learning

Matthias Weissenbacher, Rishabh Agarwal, Yoshinobu Kawahara

ICML 2024posterarXiv:2406.15025

Stochastic Q-learning for Large Discrete Action Spaces

Fares Fourati, Vaneet Aggarwal, Mohamed-Slim Alouini

ICML 2024posterarXiv:2405.10310

Successor Features for Efficient Multi-Subject Controlled Text Generation

Meng Cao, Mehdi Fatemi, Jackie Chi Kit Cheung et al.

ICML 2024poster

To the Max: Reinventing Reward in Reinforcement Learning

Grigorii Veviurko, Wendelin Boehmer, Mathijs de Weerdt

ICML 2024posterarXiv:2402.01361

Towards Optimal Adversarial Robust Q-learning with Bellman Infinity-error

Haoran Li, Zicheng Zhang, Wang Luo et al.

ICML 2024posterarXiv:2402.02165

Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning

Zhiheng Xi, Wenxiang Chen, Boyang Hong et al.

ICML 2024posterarXiv:2402.05808

ULTRAFEEDBACK: Boosting Language Models with Scaled AI Feedback

Ganqu Cui, Lifan Yuan, Ning Ding et al.

ICML 2024posterarXiv:2310.01377

Value-Evolutionary-Based Reinforcement Learning

Pengyi Li, Jianye Hao, Hongyao Tang et al.

ICML 2024oral

When Do Skills Help Reinforcement Learning? A Theoretical Analysis of Temporal Abstractions

Zhening Li, Gabriel Poesia, Armando Solar-Lezama

ICML 2024oralarXiv:2406.07897

When is Transfer Learning Possible?

My Phan, Kianté Brantley, Stephanie Milani et al.

ICML 2024poster

Zero-Shot Reinforcement Learning via Function Encoders

Tyler Ingebrand, Amy Zhang, Ufuk Topcu

ICML 2024posterarXiv:2401.17173