CMT-DeepLab: Clustering Mask Transformers for Panoptic Segmentation

Cited 25 time in webofscience Cited 0 time in scopus
  • Hit : 72
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorYu, Qihangko
dc.contributor.authorWang, Huiyuko
dc.contributor.authorKim, Dahunko
dc.contributor.authorQiao, Siyuanko
dc.contributor.authorCollins, Maxwellko
dc.contributor.authorZhu, Yukunko
dc.contributor.authorAdam, Hartwigko
dc.contributor.authorYuille, Alanko
dc.contributor.authorChen, Liang-Chiehko
dc.date.accessioned2023-09-21T01:00:31Z-
dc.date.available2023-09-21T01:00:31Z-
dc.date.created2023-09-21-
dc.date.issued2022-06-
dc.identifier.citation2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, pp.2550 - 2560-
dc.identifier.issn1063-6919-
dc.identifier.urihttp://hdl.handle.net/10203/312790-
dc.description.abstractWe propose Clustering Mask Transformer (CMT-DeepLab), a transformer-based framework for panoptic segmentation designed around clustering. It rethinks the existing transformer architectures used in segmentation and detection; CMT-DeepLab considers the object queries as cluster centers, which fill the role of grouping the pixels when applied to segmentation. The clustering is computed with an alternating procedure, by first assigning pixels to the clusters by their feature affinity, and then updating the cluster centers and pixel features. Together, these operations comprise the Clustering Mask Transformer (CMT) layer, which produces cross-attention that is denser and more consistent with the final segmentation task. CMT-DeepLab improves the performance over prior art significantly by 4.4% PQ, achieving a new state-of-the-art of 55.7% PQ on the COCO test-dev set.-
dc.languageEnglish-
dc.publisherIEEE Computer Society-
dc.titleCMT-DeepLab: Clustering Mask Transformers for Panoptic Segmentation-
dc.typeConference-
dc.identifier.wosid000867754202080-
dc.identifier.scopusid2-s2.0-85141776953-
dc.type.rimsCONF-
dc.citation.beginningpage2550-
dc.citation.endingpage2560-
dc.citation.publicationname2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationNew Orleans, LA-
dc.identifier.doi10.1109/CVPR52688.2022.00259-
dc.contributor.localauthorKim, Dahun-
dc.contributor.nonIdAuthorYu, Qihang-
dc.contributor.nonIdAuthorWang, Huiyu-
dc.contributor.nonIdAuthorQiao, Siyuan-
dc.contributor.nonIdAuthorCollins, Maxwell-
dc.contributor.nonIdAuthorZhu, Yukun-
dc.contributor.nonIdAuthorAdam, Hartwig-
dc.contributor.nonIdAuthorYuille, Alan-
dc.contributor.nonIdAuthorChen, Liang-Chieh-
Appears in Collection
RIMS Conference Papers
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 25 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0