Selective multi-cotraining for video concept detection

Niaz, Usman; Merialdo, Bernard
ICMR 2014, 4th ACM International Conference on Multimedia Retrieval, April 1-4, 2014, Glasgow, Scotland

Research interest in cotraining is increasing which combines information from usually two classi ers to iteratively increase training resources and strengthen the classi ers. We try to select classi ers for cotraining when more than two representations of the data are available. The classi er based on the selected representation or data descriptor is expected to provide the most complementary information as new labels for the target classi er. These labels are critical for the next learning iteration. We present two criteria to select the complementary classi er where classi cation results on a validation set are used to calculate statistics for all the available classi ers. These statistics are used not only to pick the best classi er but also ascertain the number of new labels to be added for the target classi er. We demonstrate the e ectiveness of classi er selection for semantic indexing task on the TRECIVD 2013 dataset and compare it to the self-training.

DOI
Type:
Conférence
City:
Glasgow
Date:
2014-04-01
Department:
Data Science
Eurecom Ref:
4270
Copyright:
© ACM, 2014. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in ICMR 2014, 4th ACM International Conference on Multimedia Retrieval, April 1-4, 2014, Glasgow, Scotland http://dx.doi.org/10.1145/2578726.2578789

PERMALINK : https://www.eurecom.fr/publication/4270