Low-level motion activity features for semantic characterization of video

K. A. Peker, A. Aydin Alatan, A. N. Akansu

Research output: Contribution to conferencePaperpeer-review

26 Scopus citations

Abstract

Efficient methods of content characterization for the browsing, retrieval or filtering of vast amount of digital video content has become a necessity. Still, there is a gap between the computationally available measures of content characteristics and the semantic interpretations of these characteristics. We want to establish connections between motion activity characteristics of video segments and the semantic characterization of them. For this purpose, two simple descriptors for motion activity of a video content is used to infer high-level semantic features of video in certain contexts. One of these descriptors, monotonous activity, is defined as the average block-based motion vector magnitude. The second descriptor, nonmonotonous activity, is an approximation to the average temporal derivative of motion vectors. Simulation results for browsing and retrieval applications show that by using the two measures together, object motions that occur close to the camera can be distinguished from distant ones. Also by using the two descriptors together, we are able to differentiate between a high activity due to camera motion and a high activity due to dancing people. Hence, these simple descriptors, especially when used to complete each other, promise to provide important clues about semantics of a video.

Original languageEnglish (US)
Pages801-804
Number of pages4
StatePublished - 2000
Event2000 IEEE International Conference on Multimedia and Expo (ICME 2000) - New York, NY, United States
Duration: Jul 30 2000Aug 2 2000

Other

Other2000 IEEE International Conference on Multimedia and Expo (ICME 2000)
Country/TerritoryUnited States
CityNew York, NY
Period7/30/008/2/00

All Science Journal Classification (ASJC) codes

  • General Engineering

Fingerprint

Dive into the research topics of 'Low-level motion activity features for semantic characterization of video'. Together they form a unique fingerprint.

Cite this