ICML 2024 "offline reinforcement learning" Papers

38 papers found

Adaptive Advantage-Guided Policy Regularization for Offline Reinforcement Learning

Tenglong Liu, Yang Li, Yixing Lan et al.

ICML 2024poster

A Primal-Dual Algorithm for Offline Constrained Reinforcement Learning with Linear MDPs

Kihyuk Hong, Ambuj Tewari

ICML 2024poster

ATraDiff: Accelerating Online Reinforcement Learning with Imaginary Trajectories

Qianlan Yang, Yu-Xiong Wang

ICML 2024poster

Bayesian Design Principles for Offline-to-Online Reinforcement Learning

Hao Hu, yiqin yang, Jianing Ye et al.

ICML 2024poster

Causal Action Influence Aware Counterfactual Data Augmentation

Núria Armengol Urpí, Marco Bagatella, Marin Vlastelica et al.

ICML 2024poster

Confidence Aware Inverse Constrained Reinforcement Learning

Sriram Ganapathi Subramanian, Guiliang Liu, Mohammed Elmahgiubi et al.

ICML 2024poster

Contrastive Representation for Data Filtering in Cross-Domain Offline Reinforcement Learning

Xiaoyu Wen, Chenjia Bai, Kang Xu et al.

ICML 2024poster

Debiased Offline Representation Learning for Fast Online Adaptation in Non-stationary Dynamics

Xinyu Zhang, Wenjie Qiu, Yi-Chen Li et al.

ICML 2024poster

DiffStitch: Boosting Offline Reinforcement Learning with Diffusion-based Trajectory Stitching

Guanghe Li, Yixiang Shan, Zhengbang Zhu et al.

ICML 2024poster

Discovering Multiple Solutions from a Single Task in Offline Reinforcement Learning

Takayuki Osa, Tatsuya Harada

ICML 2024poster

Efficient Policy Evaluation with Offline Data Informed Behavior Policy Design

Shuze Liu, Shangtong Zhang

ICML 2024poster

Enhancing Value Function Estimation through First-Order State-Action Dynamics in Offline Reinforcement Learning

Yun-Hsuan Lien, Ping-Chun Hsieh, Tzu-Mao Li et al.

ICML 2024poster

Exploration and Anti-Exploration with Distributional Random Network Distillation

Kai Yang, jian tao, Jiafei Lyu et al.

ICML 2024poster

Federated Offline Reinforcement Learning: Collaborative Single-Policy Coverage Suffices

Jiin Woo, Laixi Shi, Gauri Joshi et al.

ICML 2024poster

HarmoDT: Harmony Multi-Task Decision Transformer for Offline Reinforcement Learning

Shengchao Hu, Ziqing Fan, Li Shen et al.

ICML 2024poster

Improving Generalization in Offline Reinforcement Learning via Adversarial Data Splitting

Da Wang, Lin Li, Wei Wei et al.

ICML 2024poster

In-Context Decision Transformer: Reinforcement Learning via Hierarchical Chain-of-Thought

sili huang, Jifeng Hu, Hechang Chen et al.

ICML 2024poster

Inferring the Long-Term Causal Effects of Long-Term Treatments from Short-Term Experiments

Allen Tran, Aurelien Bibaut, Nathan Kallus

ICML 2024poster

Information-Directed Pessimism for Offline Reinforcement Learning

Alec Koppel, Sujay Bhatt, Jiacheng Guo et al.

ICML 2024poster

Is Inverse Reinforcement Learning Harder than Standard Reinforcement Learning? A Theoretical Perspective

Lei Zhao, Mengdi Wang, Yu Bai

ICML 2024poster

Learning a Diffusion Model Policy from Rewards via Q-Score Matching

Michael Psenka, Alejandro Escontrela, Pieter Abbeel et al.

ICML 2024poster

Listwise Reward Estimation for Offline Preference-based Reinforcement Learning

Heewoong Choi, Sangwon Jung, Hongjoon Ahn et al.

ICML 2024poster

Model-based Reinforcement Learning for Confounded POMDPs

Mao Hong, Zhengling Qi, Yanxun Xu

ICML 2024poster

Model-Free Robust $\phi$-Divergence Reinforcement Learning Using Both Offline and Online Data

Kishan Panaganti, Adam Wierman, Eric Mazumdar

ICML 2024poster

More Benefits of Being Distributional: Second-Order Bounds for Reinforcement Learning

Kaiwen Wang, Owen Oertell, Alekh Agarwal et al.

ICML 2024poster

Offline Actor-Critic Reinforcement Learning Scales to Large Models

Jost Tobias Springenberg, Abbas Abdolmaleki, Jingwei Zhang et al.

ICML 2024oral

Offline Transition Modeling via Contrastive Energy Learning

Ruifeng Chen, Chengxing Jia, Zefang Huang et al.

ICML 2024poster

Pessimism Meets Risk: Risk-Sensitive Offline Reinforcement Learning

Dake Zhang, Boxiang Lyu, Shuang Qiu et al.

ICML 2024spotlight

PlanDQ: Hierarchical Plan Orchestration via D-Conductor and Q-Performer

Chang Chen, Junyeob Baek, Fei Deng et al.

ICML 2024poster

Position: Reinforcement Learning in Dynamic Treatment Regimes Needs Critical Reexamination

Zhiyao Luo, Yangchen Pan, Peter Watkinson et al.

ICML 2024spotlight

Q-value Regularized Transformer for Offline Reinforcement Learning

Shengchao Hu, Ziqing Fan, Chaoqin Huang et al.

ICML 2024poster

ReDiffuser: Reliable Decision-Making Using a Diffuser with Confidence Estimation

Nantian He, Shaohui Li, Zhi Li et al.

ICML 2024poster

Reinformer: Max-Return Sequence Modeling for Offline RL

Zifeng Zhuang, Dengyun Peng, Jinxin Liu et al.

ICML 2024poster

Rethinking Decision Transformer via Hierarchical Reinforcement Learning

Yi Ma, Jianye Hao, Hebin Liang et al.

ICML 2024poster

SeMOPO: Learning High-quality Model and Policy from Low-quality Offline Visual Datasets

Shenghua Wan, Ziyuan Chen, Le Gan et al.

ICML 2024poster

Simple Ingredients for Offline Reinforcement Learning

Edoardo Cetin, Andrea Tirinzoni, Matteo Pirotta et al.

ICML 2024poster

Temporal Logic Specification-Conditioned Decision Transformer for Offline Safe Reinforcement Learning

Zijian Guo, Weichao Zhou, Wenchao Li

ICML 2024oral

Towards Robust Model-Based Reinforcement Learning Against Adversarial Corruption

Chenlu Ye, Jiafan He, Quanquan Gu et al.

ICML 2024poster