Meta-learning memory representations for continual learning with scalar-output networks스칼라 출력 신경망을 통한 연속학습에 적합한 기억 표현법 메타학습

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 149
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorSung, Youngchul-
dc.contributor.advisor성영철-
dc.contributor.authorPark, Jong Eui-
dc.date.accessioned2021-05-13T19:33:18Z-
dc.date.available2021-05-13T19:33:18Z-
dc.date.issued2020-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=911330&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/284721-
dc.description학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2020.2,[iii, 20 p. :]-
dc.description.abstractNeural networks have achieved remarkable success in various areas, but they still fail when the task changes over time. The subfield of machine learning that focuses on overcoming this problem is called continual learning. In this paper, we applied meta‐learning techniques to train a continual learning agent that utilizes an external memory via a novel scalar‐output architecture. Unlike previous works that use a fixed method to access and update memory, our agent learns the optimal way from scratch. We show that our agent successfully learned what to remember and how to recall, achieving final average accuracy about 10 % higher compared to other continual learning algorithms.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subjectContinual Learning▼aDeep Learning▼aMeta Learning▼aSequential Learning▼aOnline Learning-
dc.subject연속학습▼a심층학습▼a메타학습▼a순차학습▼a온라인학습-
dc.titleMeta-learning memory representations for continual learning with scalar-output networks-
dc.title.alternative스칼라 출력 신경망을 통한 연속학습에 적합한 기억 표현법 메타학습-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :전기및전자공학부,-
dc.contributor.alternativeauthor박종의-
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0