Enhanced Transformer Architecture for Natural Language Processing

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 143
  • Download : 0
Transformer is a state-of-the-art model in the field of natural language processing (NLP). Current NLP models primarily increase the number of transformers to improve processing performance. However, this technique requires a lot of training resources such as computing capacity. In this paper, a novel structure of Transformer is proposed. It is featured by full layer normalization, weighted residual connection, positional encoding exploiting reinforcement learning, and zero masked self-attention. The proposed Transformer model, which is called Enhanced Transformer, is validated by the bilingual evaluation understudy (BLEU) score obtained with the Multi30k translation dataset. As a result, the Enhanced Transformer achieves 202.96% higher BLEU score as compared to the original transformer with the translation dataset.
Publisher
Pacific Asia Conference on Language, Information and Computation
Issue Date
2023-12-02
Language
English
Citation

The 37th Pacific Asia Conference on Language, Information and Computation, PACLIC 37

URI
http://hdl.handle.net/10203/316729
Appears in Collection
GT-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0