Semantic concept detection for user-generated video content using a refined image folksonomy

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 381
  • Download : 3
The automatic detection of semantic concepts is a key technology for enabling efficient and effective video content management. Conventional techniques for semantic concept detection in video content still suffer from several interrelated issues: the semantic gap, the Unbalanced data set problem, and a limited concept vocabulary size. In this paper, we propose to perform semantic concept detection for user-created video content using an image folksonomy in order to overcome the aforementioned problems. First, an image folksonomy contains a vast amount of user-contributed images. Second, a significant portion of these images has been manually annotated by users using a wide variety of tags. However, user-supplied annotations in an image folksonomy are often characterized by a high level of noise. Therefore, we also discuss a method that allows reducing the number of noisy tags in an image folksonomy. This tag refinement method makes use of tag co-occurrence statistics. To verify the effectiveness of the proposed video content annotation system, experiments were performed with user-created image and video content available on a number of social media applications. For the datasets used, video annotation with tag refinement has an average recall rate of 84% and an average precision of 75%, while video annotation without tag refinement shows an average recall rate of 78% and an average precision of 62%.
Publisher
SPRINGER
Issue Date
2010-01-06
Language
English
Citation

16th International Multimedia Modeling Conference on Advances in Multimedia Modeling, MMM 2010, pp.511 - 521

DOI
10.1007/978-3-642-11301-7_51
URI
http://hdl.handle.net/10203/22820
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0