Decomposing motion and content for natural video sequence prediction

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 261
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorVillegas, Rubenko
dc.contributor.authorYang, Jimeiko
dc.contributor.authorHong, Seunghoonko
dc.contributor.authorLin, Xunyuko
dc.contributor.authorLee, Honglakko
dc.date.accessioned2020-10-23T01:57:05Z-
dc.date.available2020-10-23T01:57:05Z-
dc.date.created2020-10-06-
dc.date.issued2017-04-24-
dc.identifier.citation5th International Conference on Learning Representations, ICLR 2017-
dc.identifier.urihttp://hdl.handle.net/10203/276943-
dc.description.abstractWe propose a deep neural network for the prediction of future frames in natural video sequences. To effectively handle complex evolution of pixels in videos, we propose to decompose the motion and content, two key components generating dynamics in videos. Our model is built upon the Encoder-Decoder Convolutional Neural Network and Convolutional LSTM for pixel-level prediction, which independently capture the spatial layout of an image and the corresponding temporal dynamics. By independently modeling motion and content, predicting the next frame reduces to converting the extracted content features into the next frame content by the identified motion features, which simplifies the task of prediction. Our model is end-to-end trainable over multiple time steps, and naturally learns to decompose motion and content without separate training. We evaluate the proposed network architecture on human activity videos using KTH, Weizmann action, and UCF-101 datasets. We show state-of-the-art performance in comparison to recent approaches. To the best of our knowledge, this is the first end-to-end trainable network architecture with motion and content separation to model the spatio-temporal dynamics for pixel-level future prediction in natural videos.-
dc.languageEnglish-
dc.publisherInternational Conference on Learning Representations, ICLR-
dc.titleDecomposing motion and content for natural video sequence prediction-
dc.typeConference-
dc.identifier.scopusid2-s2.0-85064824515-
dc.type.rimsCONF-
dc.citation.publicationname5th International Conference on Learning Representations, ICLR 2017-
dc.identifier.conferencecountryFR-
dc.identifier.conferencelocationToulon-
dc.contributor.localauthorHong, Seunghoon-
dc.contributor.nonIdAuthorVillegas, Ruben-
dc.contributor.nonIdAuthorYang, Jimei-
dc.contributor.nonIdAuthorLin, Xunyu-
dc.contributor.nonIdAuthorLee, Honglak-
Appears in Collection
CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0