Modality Mixer for Multi-modal Action Recognition

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 87
  • Download : 0
In multi-modal action recognition, it is important to consider not only the complementary nature of different modalities but also global action content. In this paper, we propose a novel network, named Modality Mixer (M-Mixer) network, to leverage complementary information across modalities and temporal context of an action for multi-modal action recognition. We also introduce a simple yet effective recurrent unit, called Multi-modal Contextualization Unit (MCU), which is a core component of M-Mixer. Our MCU temporally encodes a sequence of one modality (e.g., RGB) with action content features of other modalities (e.g., depth, IR). This process encourages M-Mixer to exploit global action content and also to supplement complementary information of other modalities. As a result, our proposed method outperforms state-of-the-art methods on NTU RGB+D 60, NTU RGB+D 120, and NW-UCLA datasets. Moreover, we demonstrate the effectiveness of M-Mixer by conducting comprehensive ablation studies.
Publisher
Institute of Electrical and Electronics Engineers Inc.
Issue Date
2023-01
Language
English
Citation

23rd IEEE/CVF Winter Conference on Applications of Computer Vision, WACV 2023, pp.3297 - 3306

DOI
10.1109/WACV56688.2023.00331
URI
http://hdl.handle.net/10203/305982
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0