FaceSyncNet: A deep learning-based approach for non-linear synchronization of facial performance videos

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 50
  • Download : 0
Given a pair of facial performance videos, we present a deep learning-based approach that can automatically return a synchronized version of these videos. Traditional methods require precise facial landmark tracking and/or clean audio, and thus are sensitive to tracking inaccuracies and audio noise. To alleviate these issues, our approach leverages large-scale video datasets along with their associated audio tracks and trains a deep learning network to learn the audio descriptors of a given video frame. We then use these descriptors to compute the similarity between video frames in a cost matrix and compute a low-cost non-linear synchronization path. Both quantitative and qualitative evaluations have shown that our approach outperforms existing state-of-the-art methods.
Publisher
Institute of Electrical and Electronics Engineers Inc.
Issue Date
2019-10
Language
English
Citation

17th IEEE/CVF International Conference on Computer Vision Workshop, ICCVW 2019, pp.3703 - 3707

ISSN
2473-9936
DOI
10.1109/ICCVW.2019.00458
URI
http://hdl.handle.net/10203/311819
Appears in Collection
GCT-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0