Leveraging Large Language Models With Vocabulary Sharing For Sign Language Translation

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 35
  • Download : 0
Sign language translation (SLT) is a task that provides translation between spoken and sign languages used in the same country, which tend to show high lexical similarity but low syntactic similarity. The recent emergence of large language models (LLMs) has been remarkable for all downstream tasks in natural language processing, but they have yet to be applied to SLT. In this paper, we explore how to use an LLM with vocabulary sharing for two gloss-based SLT tasks (text-to-gloss (T2G) and gloss-to-text (G2T)) on the NIASL2021 dataset, which consists of 180,848 preprocessed Korean and Korean Sign Language (KSL) sentence pairs. The experimental results showed that Ko-GPT-Trinity-1.2B+VS, a GPT-3-based SLT model with vocabulary sharing, outperformed other SLT models, achieving BLEU-4 scores of 22.06 and 45.89 on T2G and G2T tasks, respectively. We expect that the adoption of an LLM with vocabulary sharing will significantly lessen the resource scarcity problem of SLT.
Publisher
Institute of Electrical and Electronics Engineers Inc.
Issue Date
2023-06-08
Language
English
Citation

2023 IEEE International Conference on Acoustics, Speech and Signal Processing Workshops, ICASSPW 2023

DOI
10.1109/ICASSPW59220.2023.10193533
URI
http://hdl.handle.net/10203/314596
Appears in Collection
CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0