K-mixup: Data augmentation for offline reinforcement learning using mixup in a Koopman invariant subspace

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 222
  • Download : 0
In this study, we propose a new data augmentation technique, Koopman-mixup (K-mixup), to improve the learning stability and final performance of offline reinforcement learning (RL) algorithms. K-mixup learns a Koopman invariant subspace to incorporate mixup augmentation, commonly used for classification tasks, into an RL framework. Mixup augmentation itself is known to be incompatible with RL because RL generally uses nonlinearly propagating state-based sequential inputs, whereas mixup relies on linear interpolation between a pair of inputs. To resolve the problem, Koopman embedding is used to convert a nonlinear system to a linear system, allowing successful mixup on arbitrary data pairs in any dataset. We evaluate the performance of K-mixup on several OpenAI Gym benchmark control simulations and compare it with the performance of other data augmentation methods. The comparison shows that only the proposed K-mixup consistently outperforms the base offline RL algorithm (CQL).
Publisher
PERGAMON-ELSEVIER SCIENCE LTD
Issue Date
2023-09
Language
English
Article Type
Article
Citation

EXPERT SYSTEMS WITH APPLICATIONS, v.225

ISSN
0957-4174
DOI
10.1016/j.eswa.2023.120136
URI
http://hdl.handle.net/10203/306982
Appears in Collection
ME-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0