Constrained bayesian reinforcement learning via approximate linear programming근사 선형계획법을 이용한 제약을 갖는 베이지안 강화학습

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 419
  • Download : 0
In many situations, too much exploratory behaviours can cause severe damage to the reinforcement learning agent and there should be restrictions on such behaviours. These restrictions can naturally be encoded as CMDPs where cost functions and cost constraints represent the risk of behaviours and the degree of risk taking respectively. We propose model-based Bayesian reinforcement learning (BRL) algorithm in CMDP environment, showing risk-sensitive exploration in a principled way. Our algorithm efficiently solve the given constrained BRL problem through finite approximation of the original belief-state CMDP's linear program, and generates a finite state controller in an off-line manner. We provide the corresponding theoretical guarantees and empirical supports that the proposed method outperforms the previous state-of-the-art approach.
Advisors
Kim, Kee-Eungresearcher김기응researcher
Description
한국과학기술원 :전산학부,
Publisher
한국과학기술원
Issue Date
2017
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전산학부, 2017.2,[iii, 30 p. :]

Keywords

Bayesian reinforcement learning; Safe reinforcement learning; Constrained partially observable Markov decision processes; Linear programming; 베이지안 강화 학습; 안전한 강화 학습; 비용 제약이 있는 부분 관찰 마코프 의사 결정 문제; 선형계획법

URI
http://hdl.handle.net/10203/243424
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=675468&flag=dissertation
Appears in Collection
CS-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0