Continual learning method by estimating previous task gradient via Taylor approximation이전 테스크의 테일러 근사된 경사 예측을 통한 연속학습

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 175
  • Download : 0
I address the gap between the regularization and memory based methods for mitigating catastrophic forgetting in the continual learning scenario. Memory based methods, especially experience replay(ER), store the few samples in the replay memory to use it later tasks. Many researches show that memory based approaches can easily achieve the state of the art performance, however, those methods still need to store samples from previous tasks, which violates the common assumption of continual learning problem. Although the regularization based methods can maintain this assumption, those methods still far from the state of the art performance. To mitigate this performance gap without stored samples, I study how to approximate ER methods directly, unlike the purpose of the regularization based methods. I choose the method to approximate previous task gradients by first-order Taylor approximation with the stored gradients and diagonal hessian values at final previous task parameter point. The result in the various data-sets shows that this approximation method can make better or comparable performance with the regularization-based methods, which use fixed neural network size and without stored examples.
Advisors
Chung, SaeYoungresearcher정세영researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2020
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2020.8,[iv, 25 p. :]

Keywords

Sequential Learning▼aContinual Learning▼aCatastrophic Forgetting▼aGradient Based Optimization▼aRegularization Method; 순차적 학습▼a연속 학습▼a망각▼a경사하강도법▼a정규화 방법

URI
http://hdl.handle.net/10203/285069
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=925233&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0