Large-scale meta-learning with continual trajectory shifting연속적 경로 이동을 통한 대규모 메타러닝

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 109
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorHwang, Sung Ju-
dc.contributor.advisor황성주-
dc.contributor.authorShin, Jae Woong-
dc.date.accessioned2022-04-15T07:56:33Z-
dc.date.available2022-04-15T07:56:33Z-
dc.date.issued2021-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=963741&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/294849-
dc.description학위논문(석사) - 한국과학기술원 : AI대학원, 2021.8,[iv, 31 p. :]-
dc.description.abstractMeta-learning of shared initialization parameters has shown to be highly effective in solving few-shot learning tasks. However, extending the framework to many-shot scenarios, which may further enhance its practicality, has been relatively overlooked due to the technical difficulties of meta-learning over long chains of inner-gradient steps. In this paper, we first show that allowing the meta-learners to take a larger number of inner gradient steps better captures the structure of heterogeneous and large-scale task distributions, thus results in obtaining better initialization points. Further, in order to increase the frequency of meta-updates even with the excessively long inner-optimization trajectories, we propose to estimate the $\emph{required shift}$ of the task-specific parameters with respect to the change of the initialization parameters. By doing so, we can arbitrarily increase the frequency of meta-updates and thus greatly improve the meta-level convergence as well as the quality of the learned initializations. We validate our method on a heterogeneous set of large-scale tasks and show that the algorithm largely outperforms the previous first-order meta-learning methods in terms of both generalization performance and convergence, as well as multi-task learning and fine-tuning baselines.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subjectMeta learning▼aLarge scale-
dc.subject메타 러닝▼a대규모 학습-
dc.titleLarge-scale meta-learning with continual trajectory shifting-
dc.title.alternative연속적 경로 이동을 통한 대규모 메타러닝-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :AI대학원,-
dc.contributor.alternativeauthor신재웅-
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0