Marker-Embedded Tactile Image Generation via Generative Adversarial Networks

Cited 1 time in webofscience Cited 0 time in scopus
  • Hit : 92
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorKim, Won Dongko
dc.contributor.authorYang, Sanghoonko
dc.contributor.authorKim, Woojongko
dc.contributor.authorKim, Jeong-Jungko
dc.contributor.authorKim, Chang-Hyunko
dc.contributor.authorKim, Jungko
dc.date.accessioned2023-07-13T02:00:28Z-
dc.date.available2023-07-13T02:00:28Z-
dc.date.created2023-07-13-
dc.date.created2023-07-13-
dc.date.issued2023-08-
dc.identifier.citationIEEE ROBOTICS AND AUTOMATION LETTERS, v.8, no.8, pp.4481 - 4488-
dc.identifier.issn2377-3766-
dc.identifier.urihttp://hdl.handle.net/10203/310470-
dc.description.abstractData-driven methods have been successfully applied to images from vision-based tactile sensors to fulfill various manipulation tasks. Nevertheless, these methods remain inefficient because of the lack of methods for simulating the sensors. Relevant research on simulating vision-based tactile sensors generally focus on generating images without markers, owing to the challenges in accurately generating marker motions caused by elastomer deformation. This disallows access to tactile information deducible from markers. In this letter, we propose a generative adversarial network (GAN)-based method to generate realistic marker-embedded tactile images in GelSight-like vision-based tactile sensors. We trained the proposed GAN model with an aligned real tactile and simulated depth image dataset obtained from deforming the sensor against various objects. This allows the model to translate simulated depth image sequences into RGB tactile images with markers. Furthermore, the generator in the proposed GAN allows the network to integrate the history of deformations from the depth image sequences to generate realistic marker motions during the normal and lateral sensor deformations. We evaluated and compared the positional accuracy of the markers and image similarity metrics of the images generated via our method with those from prior methods. The generated tactile images from the proposed model show a 28.3% decrease in marker positional error and a 93.5% decrease in the image similarity metric (MSE) compared to those generated by previous methods, validating the effectiveness of our approach.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleMarker-Embedded Tactile Image Generation via Generative Adversarial Networks-
dc.typeArticle-
dc.identifier.wosid001012677300007-
dc.identifier.scopusid2-s2.0-85162722836-
dc.type.rimsART-
dc.citation.volume8-
dc.citation.issue8-
dc.citation.beginningpage4481-
dc.citation.endingpage4488-
dc.citation.publicationnameIEEE ROBOTICS AND AUTOMATION LETTERS-
dc.identifier.doi10.1109/LRA.2023.3284370-
dc.contributor.localauthorKim, Jung-
dc.contributor.nonIdAuthorYang, Sanghoon-
dc.contributor.nonIdAuthorKim, Jeong-Jung-
dc.contributor.nonIdAuthorKim, Chang-Hyun-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorDeep learning methods-
dc.subject.keywordAuthorforce and tactile sensing-
dc.subject.keywordAuthorsimulation and animation-
dc.subject.keywordPlusTO-REAL TRANSFER-
dc.subject.keywordPlusDOMAIN ADAPTATION-
dc.subject.keywordPlusPERCEPTION-
dc.subject.keywordPlusSENSORS-
Appears in Collection
ME-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 1 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0