SURF: Semi-supervised Reward Learning with Data Augmentation for Feedback-efficient Preference-based Reinforcement Learning

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 75
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorPark, Jongjinko
dc.contributor.authorSeo, Younggyoko
dc.contributor.authorShin, Jinwooko
dc.contributor.authorLee, Honglakko
dc.contributor.authorAbbeel, Pieterko
dc.contributor.authorLee, Kiminko
dc.date.accessioned2023-12-12T11:03:00Z-
dc.date.available2023-12-12T11:03:00Z-
dc.date.created2023-12-08-
dc.date.created2023-12-08-
dc.date.issued2022-04-26-
dc.identifier.citation10th International Conference on Learning Representations, ICLR 2022-
dc.identifier.urihttp://hdl.handle.net/10203/316334-
dc.description.abstractPreference-based reinforcement learning (RL) has shown potential for teaching agents to perform the target tasks without a costly, pre-defined reward function by learning the reward with a supervisor's preference between the two agent behaviors. However, preference-based learning often requires a large amount of human feedback, making it difficult to apply this approach to various applications. This data-efficiency problem, on the other hand, has been typically addressed by using unlabeled samples or data augmentation techniques in the context of supervised learning. Motivated by the recent success of these approaches, we present SURF, a semi-supervised reward learning framework that utilizes a large amount of unlabeled samples with data augmentation. In order to leverage unlabeled samples for reward learning, we infer pseudo-labels of the unlabeled samples based on the confidence of the preference predictor. To further improve the label-efficiency of reward learning, we introduce a new data augmentation that temporally crops consecutive subsequences from the original behaviors. Our experiments demonstrate that our approach significantly improves the feedback-efficiency of the state-of-the-art preference-based method on a variety of locomotion and robotic manipulation tasks.-
dc.languageEnglish-
dc.publisherInternational Conference on Learning Representations-
dc.titleSURF: Semi-supervised Reward Learning with Data Augmentation for Feedback-efficient Preference-based Reinforcement Learning-
dc.typeConference-
dc.identifier.scopusid2-s2.0-85140724160-
dc.type.rimsCONF-
dc.citation.publicationname10th International Conference on Learning Representations, ICLR 2022-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationVirtual-
dc.contributor.localauthorShin, Jinwoo-
dc.contributor.localauthorLee, Kimin-
dc.contributor.nonIdAuthorLee, Honglak-
dc.contributor.nonIdAuthorAbbeel, Pieter-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0