Large-Scale Meta-Learning with Continual Trajectory Shifting

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 314
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorShin, Jae Woongko
dc.contributor.authorLee, Hae Beomko
dc.contributor.authorGong, Boqingko
dc.contributor.authorHwang, Sung Juko
dc.date.accessioned2021-12-10T06:49:16Z-
dc.date.available2021-12-10T06:49:16Z-
dc.date.created2021-11-30-
dc.date.created2021-11-30-
dc.date.created2021-11-30-
dc.date.issued2021-07-18-
dc.identifier.citation38th International Conference on Machine Learning, ICML 2021-
dc.identifier.issn2640-3498-
dc.identifier.urihttp://hdl.handle.net/10203/290420-
dc.description.abstractMeta-learning of shared initialization parameters has shown to be highly effective in solving few-shot learning tasks. However, extending the framework to many-shot scenarios, which may further enhance its practicality, has been relatively overlooked due to the technical difficulties of meta-learning over long chains of inner-gradient steps. In this paper, we first show that allowing the meta-learners to take a larger number of inner gradient steps better captures the structure of heterogeneous and large-scale task distributions, thus results in obtaining better initialization points. Further, in order to increase the frequency of meta-updates even with the excessively long inner-optimization trajectories, we propose to estimate the required shift of the task-specific parameters with respect to the change of the initialization parameters. By doing so, we can arbitrarily increase the frequency of meta-updates and thus greatly improve the meta-level convergence as well as the quality of the learned initializations. We validate our method on a heterogeneous set of large-scale tasks, and show that the algorithm largely outperforms the previous first-order meta-learning methods in terms of both generalization performance and convergence, as well as multi-task learning and fine-tuning baselines.-
dc.languageEnglish-
dc.publisherInternational Machine Learning Society-
dc.titleLarge-Scale Meta-Learning with Continual Trajectory Shifting-
dc.typeConference-
dc.identifier.wosid000768182705069-
dc.identifier.scopusid2-s2.0-85161272554-
dc.type.rimsCONF-
dc.citation.publicationname38th International Conference on Machine Learning, ICML 2021-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationVirtual-
dc.contributor.localauthorHwang, Sung Ju-
dc.contributor.nonIdAuthorShin, Jae Woong-
dc.contributor.nonIdAuthorGong, Boqing-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0