R-LSTM : restoring relevant contexts for better understanding of natural languageR-LSTM : 자연어 이해를 향상시키기 위해 연관 문맥을 복원하는 방법

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 509
  • Download : 0
In Natural Language Understanding (NLU), sequence models such as Recurrent Neural Network (RNN), Long Short-Term Memory (LSTM) are widely used because they can learn phrase and sentence representations that are sensitive to word order. However, the sequence models share a common limitation that previous contextual information tends to be forgotten as new words are processed. It makes the sequence models hard to learn the relations among new word and words processed in the previous. To address the limitation, we propose a sequence model called Restorable Long Short-Term Memory (R-LSTM) that can restore the forgotten information needed to understand the current word. R-LSTM is an extension of the LSTM by applying two approaches: a multiple attention mechanism with coverage loss and a restore gate. The multiple attention mechanism with coverage loss extracts the relevant contexts from the stored previous contexts. The relevant contexts are selectively reflected into current memory of R-LSTM through the restore gate. In other words, the restore gate enables our R-LSTM to restore only the forgotten and meaningful information. We evaluate our model on various NLU tasks and data sets: language modeling (WikiText-2 data set); natural language inference (SNLI data set); text classification (AG's News, Yahoo! Answers, Yelp Review Polarity data sets). In general, experimental results on various tasks show that the performance of our model outperformed in comparison of conventional sequence models. Moreover, the exploratory experiments on WikiText-2 data set are conducted to assess that each proposed approach affects positively on the performance.
Advisors
Choi, Ho-Jinresearcher최호진researcher
Description
한국과학기술원 :전산학부,
Publisher
한국과학기술원
Issue Date
2019
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전산학부, 2019.2,[iv, 34 p. :]

Keywords

Restore gate▼acoverage loss▼amultiple attention▼asequence model; 복원 게이트▼a커버리지 손실▼a다중 어탠션▼a시퀀스 모델

URI
http://hdl.handle.net/10203/267109
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=843516&flag=dissertation
Appears in Collection
CS-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0