Concept detector refinement using social videos

Liu, Xueliang; Huet, Benoit
VLS-MCMR 2010, International Workshop on Very-Large-Scale Multimedia Corpus, Mining and Retrieval, 29 October 2010, Florence, Italy

The explosion of social video sharing sites gives new challenges on video search and indexing techniques. Because of the concept diversity in social videos, it is very hard to build a well annotated dataset that provides good coverage over the whole meaning of concepts. However, the prosperity of social videos on the internet also make it easy to obtain a huge number of videos, which gives an opportunity to mine the semantic content from an infinite amount of video entities. In this paper, we focus on improving the performance concept detectors and propose a refinement framework based on a semi-supervised learning technique. In our framework, the self-training algorithm is employed to expand the training dataset with automatically labeled data. The contribution of this paper is to demonstrate how to utilize the visual feature and text metadata to enhance the performance of concept classifier with a lot number of unlabeled videos. By experimenting on a social video dataset with 21,000 entities, it is shown that after expanding the training set with automatically labeled shots, the concept detectors' performance can be significantly improved.


DOI
Type:
Conference
City:
Firenze
Date:
2010-10-29
Department:
Data Science
Eurecom Ref:
3285
Copyright:
© ACM, 2010. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in VLS-MCMR 2010, International Workshop on Very-Large-Scale Multimedia Corpus, Mining and Retrieval, 29 October 2010, Florence, Italy
http://dx.doi.org/10.1145/1878137.1878142
See also:

PERMALINK : https://www.eurecom.fr/publication/3285