Citation
Abstract
In video search and retrieval, user’s need is expressed in terms of query. Early video retrieval systems usually matched video clips with such low-level features as color, shape, texture, and motion. In spite of the fact that retrieval is done accurately and automatically with such low-level features, the semantic meaning of the query cannot be expressed in this way. Moreover, the limitation of retrieval using desirable concept detectors is providing annotations for each concept. However, providing annotation for every concept in real world is very challenging and time consuming, and it is not possible to provide annotation for every concept in the real world. In this study, in order to improve the effectiveness of the retrieval, a method for similarity computation is proposed and experimented for mapping concepts whose annotations are not available onto the annotated and known concepts. The TRECVID 2005 data set is used to evaluate the effectiveness of the concept-based video retrieval model by applying the proposed similarity method. Results are also compared with previous similarity measures used in the same domain. The proposed similarity measure approach outperforms other methods with the Mean Average Precision (MAP) of 26.84% in concept retrieval.
Download File
Full text not available from this repository.
|
Additional Metadata
Item Type: | Book Section |
---|---|
Divisions: | Faculty of Computer Science and Information Technology |
DOI Number: | https://doi.org/10.1007/978-3-642-32826-8_27 |
Publisher: | Springer |
Keywords: | Video retrieval; Video analysis; Semantic knowledge; Similarity measures |
Depositing User: | Nurul Ainie Mokhtar |
Date Deposited: | 19 Jan 2016 04:29 |
Last Modified: | 19 Jan 2016 04:29 |
Altmetrics: | http://www.altmetric.com/details.php?domain=psasir.upm.edu.my&doi=10.1007/978-3-642-32826-8_27 |
URI: | http://psasir.upm.edu.my/id/eprint/26091 |
Statistic Details: | View Download Statistic |
Actions (login required)
View Item |