Multi-level fusion for semantic video content indexing and retrieval

Benmokhtar, Rachid; Huet, Benoit
AMR 2007, International Workshop on Adaptive Multimedia Retrieval, June 5-6 2007, Paris, France | Also published as LNCS Volume 4918

In this paper, we present the results of our work on the analysis of
an automatic semantic video content indexing and retrieval system based on fusing
various low level visual and edges descriptors. Global MPEG-7 features, extracted
from video shots, are described via IVSM signature (Image Vector Space
Model) in order to have a compact description of the content. Both static and dynamic
feature fusion are introduced to obtain effective signatures. Support Vector
Machines (SVMs) are employed to perform classification (One classifier per
feature). The task of the classifiers is to detect the video semantic content. Then,
classifier outputs are fused using a neural network based on evidence theory (NNET)
in order to provide a decision on the content of each shot. The experimental
results are conducted in the framework of the TrecVid feature extraction task.


DOI
Type:
Conference
City:
Paris
Date:
2007-06-05
Department:
Data Science
Eurecom Ref:
2292
Copyright:
© Springer. Personal use of this material is permitted. The definitive version of this paper was published in AMR 2007, International Workshop on Adaptive Multimedia Retrieval, June 5-6 2007, Paris, France | Also published as LNCS Volume 4918 and is available at : http://dx.doi.org/10.1007/978-3-540-79860-6_13
See also:

PERMALINK : https://www.eurecom.fr/publication/2292