EvDistill: Asynchronous Events to End-task Learning via Bidirectional Reconstruction-guided Cross-modal Knowledge Distillation

Cited 20 time in webofscience Cited 0 time in scopus
  • Hit : 260
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorWang, Linko
dc.contributor.authorYoon, Kuk-Jinko
dc.contributor.authorChae, Yujeongko
dc.contributor.authorYoon, SungHoonko
dc.contributor.authorKim, Tae-Kyunko
dc.date.accessioned2021-07-02T02:10:19Z-
dc.date.available2021-07-02T02:10:19Z-
dc.date.created2021-04-21-
dc.date.created2021-04-21-
dc.date.created2021-04-21-
dc.date.created2021-04-21-
dc.date.created2021-04-21-
dc.date.created2021-04-21-
dc.date.issued2021-06-19-
dc.identifier.citationIEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.608 - 619-
dc.identifier.issn1063-6919-
dc.identifier.urihttp://hdl.handle.net/10203/286364-
dc.description.abstractEvent cameras sense per-pixel intensity changes and produce asynchronous event streams with high dynamic range and less motion blur, showing advantages over the conventional cameras. A hurdle of training event-based models is the lack of large qualitative labeled data. Prior works learning end-tasks mostly rely on labeled or pseudolabeled datasets obtained from the active pixel sensor (APS) frames; however, such datasets' quality is far from rivaling those based on the canonical images. In this paper, we propose a novel approach, called EvDistill, to learn a student network on the unlabeled and unpaired event data (target modality) via knowledge distillation (KD) from a teacher network trained with large-scale, labeled image data (source modality). To enable KD across the unpaired modalities, we first propose a bidirectional modality reconstruction (BMR) module to bridge both modalities and simultaneously exploit them to distill knowledge via the crafted pairs, causing no extra computation in the inference. The BMR is improved by the end-tasks and KD losses in an end-to-end manner. Second, we leverage the structural similarities of both modalities and adapt the knowledge by matching their distributions. Moreover, as most prior feature KD methods are uni-modality and less applicable to our problem, we propose an affinity graph KD loss to boost the distillation. Our extensive experiments on semantic segmentation and object recognition demonstrate that EvDistill achieves significantly better results than the prior works and KD with only events and APS frames.-
dc.languageEnglish-
dc.publisherComputer Vision Foundation, IEEE Computer Society-
dc.titleEvDistill: Asynchronous Events to End-task Learning via Bidirectional Reconstruction-guided Cross-modal Knowledge Distillation-
dc.typeConference-
dc.identifier.wosid000739917300058-
dc.type.rimsCONF-
dc.citation.beginningpage608-
dc.citation.endingpage619-
dc.citation.publicationnameIEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationVirtual-
dc.identifier.doi10.1109/CVPR46437.2021.00067-
dc.contributor.localauthorYoon, Kuk-Jin-
dc.contributor.localauthorKim, Tae-Kyun-
Appears in Collection
CS-Conference Papers(학술회의논문)ME-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 20 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0