Continual Learners are Incremental Model Generalizers

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 34
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorYoon, Jaehongko
dc.contributor.authorHwang, Sung Juko
dc.contributor.authorCao, Yueko
dc.date.accessioned2023-12-11T03:04:05Z-
dc.date.available2023-12-11T03:04:05Z-
dc.date.created2023-12-09-
dc.date.issued2023-07-26-
dc.identifier.citation40th International Conference on Machine Learning (ICML 2023)-
dc.identifier.urihttp://hdl.handle.net/10203/316204-
dc.description.abstractMotivated by the efficiency and rapid convergence of pre-trained models for solving downstream tasks, this paper extensively studies the impact of Continual Learning (CL) models as pre-trainers. We find that, in both supervised and unsupervised CL, the transfer quality of representations does not show a noticeable degradation of fine-tuning performance but rather increases gradually. This is because CL models can learn improved task-general features when easily forgetting task-specific knowledge. Based on this observation, we suggest a new unsupervised CL framework with masked modeling, which aims to capture fluent task-generic representation during training. Furthermore, we propose a new fine-tuning scheme, GLobal Attention Discretization (GLAD), that preserves rich task-generic representation during solving downstream tasks. The model fine-tuned with GLAD achieves competitive performance and can also be used as a good pre-trained model itself. We believe this paper breaks the barriers between pre-training and fine-tuning steps and leads to a sustainable learning framework in which the continual learner incrementally improves model generalization, yielding better transfer to unseen tasks.-
dc.languageEnglish-
dc.publisherInternational Machine Learning Society-
dc.titleContinual Learners are Incremental Model Generalizers-
dc.typeConference-
dc.identifier.scopusid2-s2.0-85174425945-
dc.type.rimsCONF-
dc.citation.publicationname40th International Conference on Machine Learning (ICML 2023)-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationHonolulu, HI-
dc.contributor.localauthorHwang, Sung Ju-
dc.contributor.nonIdAuthorCao, Yue-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0