Mitigating language-dependent ethnic bias in BERTBERT의 민족적 선입견에 대한 분석 및 해결 방안

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 60
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorOh, Hae Yun-
dc.contributor.advisor오혜연-
dc.contributor.authorAhn, Jaimeen-
dc.date.accessioned2023-06-26T19:31:34Z-
dc.date.available2023-06-26T19:31:34Z-
dc.date.issued2022-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=997583&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/309551-
dc.description학위논문(석사) - 한국과학기술원 : 전산학부, 2022.2,[iv, 28 p. :]-
dc.description.abstractBERT and other large-scale language models (LMs) contain gender and racial bias. They also exhibit other dimensions of social bias, most of which have not been studied in depth, and some of which vary depending on the language. In this paper, we study ethnic bias and how it varies across languages by analyzing and mitigating ethnic bias in monolingual BERT for English, German, Spanish, Korean, Turkish, and Chinese. To observe and quantify ethnic bias, we develop a novel metric called Categorical Bias score. Then we propose two methods for mitigation-
dc.description.abstractfirst using a multilingual model, and second using contextual word alignment of two monolingual models. We compare our proposed methods with monolingual BERT and show that these methods effectively alleviate the ethnic bias. Which of the two methods works better depends on the amount of NLP resources available for that language. We additionally experiment with Arabic and Greek to verify that our proposed methods work for a wider variety of languages.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.titleMitigating language-dependent ethnic bias in BERT-
dc.title.alternativeBERT의 민족적 선입견에 대한 분석 및 해결 방안-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :전산학부,-
dc.contributor.alternativeauthor안재민-
Appears in Collection
CS-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0