Moor: Model-based offline policy optimization with a risk dynamics model
Abstract Offline reinforcement learning (RL) has been widely used in safety-critical domains by avoiding dangerous and costly online interaction. A significant challenge is addressing uncertainties and risks outside of offline data. Risk-sensitive offline RL attempts to solve this issue by risk aver...
Saved in:
Main Authors: | Xiaolong Su, Peng Li, Shaofei Chen |
---|---|
Format: | Article |
Language: | English |
Published: |
Springer
2024-11-01
|
Series: | Complex & Intelligent Systems |
Subjects: | |
Online Access: | https://doi.org/10.1007/s40747-024-01621-x |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Offline prompt reinforcement learning method based on feature extraction
by: Tianlei Yao, et al.
Published: (2025-01-01) -
Stealthy data poisoning attack method on offline reinforcement learning in unmanned systems
by: ZHOU Xue, et al.
Published: (2024-12-01) -
Reinforcement Learning-Based Autonomous Soccer Agents: A Study in Multi-Agent Coordination and Strategy Development
by: Biplov Paneru, et al.
Published: (2025-01-01) -
ANALYSIS OF THE FACTORS INFLUENCING THE OFFLINE LEARNING READINESS DURING THE COVID-19 PANDEMIC
by: Arlina Dhian Sulistyowati, et al.
Published: (2023-05-01) -
Reinforcement learning algorithm based on minimum state method and average reward
by: LIU Quan1, et al.
Published: (2011-01-01)