Convolutional recurrent neural networks for first-person action recognition일인칭 행동 인식을 위한 컨볼루션 순환 신경망 연구

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 440
  • Download : 0
In recent years, research on first-person images has become increasingly important in the field of computer vision due to the development of wearable cameras and the interest in life logging. However, it is difficult to analyze the first-person image because the user 's hand is represented in various ways as well as the camera motion is mixed. As a general approach, Convolutional Neural Network (CNN) based learning methods are used primarily for vision tasks such as classification and recognition, because they better represent the latent features of an image. However, for vision work involving video data, the CNN-based model has the disadvantage that it is difficult to learn the long-time dependence between sequence data. In order to overcome such limitations, we propose a deep network structure consisting of CNN and LSTM (Long short term memory) for action recognition in first-person image data. Our model has two main concepts: First, each object information and motion information is learned through a convolution network divided into two streams. The next step is to learn the temporal dependence of multi-task learning in the LSTM model through the latent features obtained from each stream. We evaluated the performance of the GTEA dataset and compared it with other studies.
Advisors
Jo, Sunghoresearcher조성호researcher
Description
한국과학기술원 :전산학부,
Publisher
한국과학기술원
Issue Date
2017
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전산학부, 2017.2,[iii, 25 p. :]

Keywords

First-person video; Action recognition; CNN; LSTM; Multi-task learning; 일인칭 영상; 행동 인지; 멀티 태스크 학습

URI
http://hdl.handle.net/10203/243415
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=675478&flag=dissertation
Appears in Collection
CS-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0