Motion activity based semantic video similarity retrieval

3Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Semantic feature extraction of video shots and fast video sequence matching are important and required for efficient retrieval in a large video database. In this paper, a novel mechanism of similarity retrieval is proposed. Similarity measure between video sequences considering the spatio-temporal variation through consecutive frames is presented. For bridging the semantic gap between low-level features and the rich meaning that users desire to capture, video shots are analyzed and characterized by the high-level feature of motion activity in compressed domain. The extracted features of motion activity are further described by the 2D-histogram that is sensitive to the spatiotemporal variation of moving objects. In order to reduce the dimensions of feature vector space in sequence matching, Discrete Cosine Transform (DCT) is exploited to map semantic features of consecutive frames to the frequency domain while retains the discriminatory information and preserves the Euclidean distance between feature vectors. Experiments are performed on MPEG-7 testing videos, and the results of sequence matching show that a few DCT transformed coefficients are adequate and thus reveal the effectiveness of the proposed mechanism of video retrieval.

Cite

CITATION STYLE

APA

Chen, D. Y., Lee, S. Y., & Chen, H. T. (2002). Motion activity based semantic video similarity retrieval. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 2532, pp. 319–327). Springer Verlag. https://doi.org/10.1007/3-540-36228-2_40

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free