TemporalWiki: A Lifelong Benchmark for Training and Evaluating Ever-Evolving Language Models

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 37
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorJang, Joelko
dc.contributor.authorYe, Seonghyeonko
dc.contributor.authorLee, Changhoko
dc.contributor.authorYang, Soheeko
dc.contributor.authorShin, Joongboko
dc.contributor.authorHan, Janghoonko
dc.contributor.authorKim, Gyeonghunko
dc.contributor.authorSeo, Minjoonko
dc.date.accessioned2023-12-12T03:00:39Z-
dc.date.available2023-12-12T03:00:39Z-
dc.date.created2023-12-09-
dc.date.created2023-12-09-
dc.date.issued2022-12-
dc.identifier.citationEMNLP 2022, pp.6237 - 6250-
dc.identifier.urihttp://hdl.handle.net/10203/316275-
dc.description.abstractLanguage Models (LMs) become outdated as the world changes; they often fail to perform tasks requiring recent factual information which was absent or different during training, a phenomenon called temporal misalignment. This is especially a challenging problem because the research community still lacks a coherent dataset for assessing the adaptability of LMs to frequently-updated knowledge corpus such as Wikipedia. To this end, we introduce TEMPORALWIKI, a lifelong benchmark for ever-evolving LMs that utilizes the difference between consecutive snapshots of English Wikipedia and English Wikidata for training and evaluation, respectively. The benchmark hence allows researchers to periodically track an LM's ability to retain previous knowledge and acquire updated/new knowledge at each point in time. We also find that training an LM on the diff data through continual learning methods achieves similar or better perplexity than on the entire snapshot in our benchmark with 12 times less computational cost, which verifies that factual knowledge in LMs can be safely updated with minimal training data via continual learning. The dataset and the code is made available at this link.-
dc.languageEnglish-
dc.publisherAssociation for Computational Linguistics (ACL)-
dc.titleTemporalWiki: A Lifelong Benchmark for Training and Evaluating Ever-Evolving Language Models-
dc.typeConference-
dc.identifier.scopusid2-s2.0-85144265982-
dc.type.rimsCONF-
dc.citation.beginningpage6237-
dc.citation.endingpage6250-
dc.citation.publicationnameEMNLP 2022-
dc.identifier.conferencecountryAR-
dc.identifier.conferencelocationAbu Dhabi-
dc.contributor.localauthorSeo, Minjoon-
dc.contributor.nonIdAuthorJang, Joel-
dc.contributor.nonIdAuthorYe, Seonghyeon-
dc.contributor.nonIdAuthorLee, Changho-
dc.contributor.nonIdAuthorYang, Sohee-
dc.contributor.nonIdAuthorShin, Joongbo-
dc.contributor.nonIdAuthorHan, Janghoon-
dc.contributor.nonIdAuthorKim, Gyeonghun-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0