Textual Backdoor Attack for the Text Classification System

Cited 6 time in webofscience Cited 0 time in scopus
  • Hit : 130
  • Download : 59
DC FieldValueLanguage
dc.contributor.authorKwon, Hyunko
dc.contributor.authorLee, Sanghyunko
dc.date.accessioned2021-11-17T06:42:09Z-
dc.date.available2021-11-17T06:42:09Z-
dc.date.created2021-11-16-
dc.date.created2021-11-16-
dc.date.created2021-11-16-
dc.date.issued2021-10-
dc.identifier.citationSECURITY AND COMMUNICATION NETWORKS, v.2021-
dc.identifier.issn1939-0114-
dc.identifier.urihttp://hdl.handle.net/10203/289207-
dc.description.abstractDeep neural networks provide good performance for image recognition, speech recognition, text recognition, and pattern recognition. However, such networks are vulnerable to backdoor attacks. In a backdoor attack, normal data that do not include a specific trigger are correctly classified by the target model, but backdoor data that include the trigger are incorrectly classified by the target model. One advantage of a backdoor attack is that the attacker can use a specific trigger to attack at a desired time. In this study, we propose a backdoor attack targeting the BERT model, which is a classification system designed for use in the text domain. Under the proposed method, the model is additionally trained on a backdoor sentence that includes a specific trigger, and afterward, if the trigger is attached before or after an original sentence, it will be misclassified by the model. In our experimental evaluation, we used two movie review datasets (MR and IMDB). The results show that using the trigger word "ATTACK" at the beginning of an original sentence, the proposed backdoor method had a 100% attack success rate when approximately 1.0% and 0.9% of the training data consisted of backdoor samples, and it allowed the model to maintain an accuracy of 86.88% and 90.80% on the original samples in the MR and IMDB datasets, respectively.</p>-
dc.languageEnglish-
dc.publisherWILEY-HINDAWI-
dc.titleTextual Backdoor Attack for the Text Classification System-
dc.typeArticle-
dc.identifier.wosid000715784500001-
dc.identifier.scopusid2-s2.0-85118566512-
dc.type.rimsART-
dc.citation.volume2021-
dc.citation.publicationnameSECURITY AND COMMUNICATION NETWORKS-
dc.identifier.doi10.1155/2021/2938386-
dc.contributor.localauthorLee, Sanghyun-
dc.contributor.nonIdAuthorKwon, Hyun-
dc.description.isOpenAccessY-
dc.type.journalArticleArticle-
dc.subject.keywordPlusDEEP NEURAL-NETWORKS-
Appears in Collection
RIMS Journal Papers
Files in This Item
122359.pdf(2.8 MB)Download
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 6 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0