Extracting TTS style via adversarial and supervised contrastive learning적대적 및 지도 대조 학습을 통한 TTS 스타일 추출

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 101
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorYang, Eunho-
dc.contributor.advisor양은호-
dc.contributor.authorShin, Eunhyouk-
dc.date.accessioned2023-06-22T19:31:14Z-
dc.date.available2023-06-22T19:31:14Z-
dc.date.issued2022-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=997677&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/308187-
dc.description학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2022.2,[iii, 21 p. :]-
dc.description.abstractFew-shot TTS is a useful but challenging task where we have to mimic a new style given a short reference speech. A popular approach for tackling this problem is to rely on architecture bottleneck for extracting style embedding. However, this approach may have robustness issues if the extracted embedding is not independent of text input, or relevance to speaker identity might be limited due to the bottleneck. In this study, we propose to use adversarial contrastive learning to extract style independent of text. Furthermore, we propose to use supervised contrastive learning to reinforce relevance to speaker identity and utilize rich representation learned by contrastive learning. Quantitative evaluation on benchmark dataset is performed in order to show that our method indeed improves robustness and relevance to speaker identity.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.titleExtracting TTS style via adversarial and supervised contrastive learning-
dc.title.alternative적대적 및 지도 대조 학습을 통한 TTS 스타일 추출-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :김재철AI대학원,-
dc.contributor.alternativeauthor신은혁-
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0