Not Just Compete, but Collaborate: Local Image-to-Image Translation via Cooperative Mask Prediction

Cited 3 time in webofscience Cited 0 time in scopus
  • Hit : 93
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorKi, Daejinko
dc.contributor.authorKhan, Mohammad Azamko
dc.contributor.authorChoo, Jaegulko
dc.date.accessioned2021-12-10T06:49:38Z-
dc.date.available2021-12-10T06:49:38Z-
dc.date.created2021-12-03-
dc.date.created2021-12-03-
dc.date.created2021-12-03-
dc.date.created2021-12-03-
dc.date.issued2021-06-19-
dc.identifier.citationIEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.6505 - 6514-
dc.identifier.issn1063-6919-
dc.identifier.urihttp://hdl.handle.net/10203/290428-
dc.description.abstractFacial attribute editing aims to manipulate the image with the desired attribute while preserving the other details. Recently, generative adversarial networks along with the encoder-decoder architecture have been utilized for this task owing to their ability to create realistic images. However, the existing methods for the unpaired dataset cannot still preserve the attribute-irrelevant regions properly due to the absence of the ground truth image. This work proposes a novel, intuitive loss function called the CAM-consistency loss, which improves the consistency of an input image in image translation. While the existing cycle-consistency loss ensures that the image can be translated back, our approach makes the model further preserve the attribute-irrelevant regions even in a single translation to another domain by using the Grad-CAM output computed from the discriminator. Our CAM-consistency loss directly optimizes such a Grad-CAM output from the discriminator during training, in order to properly capture which local regions the generator should change while keeping the other regions unchanged. In this manner, our approach allows the generator and the discriminator to collaborate with each other to improve the image translation quality. In our experiments, we validate the effectiveness and versatility of our proposed CAM-consistency loss by applying it to several representative models for facial image editing, such as StarGAN, AttGAN, and STGAN.-
dc.languageEnglish-
dc.publisherIEEE Computer Vision and Pattern Recognition-
dc.titleNot Just Compete, but Collaborate: Local Image-to-Image Translation via Cooperative Mask Prediction-
dc.typeConference-
dc.identifier.wosid000739917306071-
dc.type.rimsCONF-
dc.citation.beginningpage6505-
dc.citation.endingpage6514-
dc.citation.publicationnameIEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationVirtual-
dc.identifier.doi10.1109/CVPR46437.2021.00644-
dc.contributor.localauthorChoo, Jaegul-
dc.contributor.nonIdAuthorKi, Daejin-
dc.contributor.nonIdAuthorKhan, Mohammad Azam-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 3 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0