Fast and Memory-Efficient Uncertainty-Aware Framework for Offline Reinforcement Learning with Rank One MIMO Q network

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 84
  • Download : 0
Offline reinforcement learning (RL) has recently gained attention for its ability to train policies from existing datasets, eliminating the need for additional environment exploration. However, applying conventional RL algorithms directly to offline RL faces a challenge due to extrapolation errors from out-of-distribution (OOD) data. Existing methods often suffer from issues such as over-conservative value function, imprecise OOD data handling, and high computational costs. To address these limitations, we introduce FMU, a framework that stands for Fast and Memory-efficient Uncertainty quantification. FMU utilizes uncertainty quantification to effectively leverage reliable OOD data while reducing conservatism for in-distribution data. Our framework precisely learns a policy by maximizing the lower confidence bound of the corresponding Q-function and employs a Rank One Multi-Input Multi-Output (MIMO) architecture to effectively make the Q-function uncertainty aware.Extensive experiments on the D4RL benchmark show that FMU achieves state-of-the-art performance while remaining computationally efficient. By incorporating uncertainty quantification,FMU offers a promising solution to mitigate extrapolation errors and enhance offline RL efficiency.
Publisher
IROS 2023 Workshop
Issue Date
2023-10
Language
English
Citation

IROS 2023 Workshop on Policy Learning in Geometric Spaces

URI
http://hdl.handle.net/10203/314971
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0