Article ID | Journal | Published Year | Pages | File Type |
---|---|---|---|---|
10369762 | Signal Processing | 2005 | 27 Pages |
Abstract
The use of video and audio features for automated annotation of audio-visual data is becoming widespread. A major limitation of many of the current methods is that the stored indexing features are too low-level-they relate directly to properties of the data. In this work we apply a further stage of processing that associates the feature measurements with real-world objects or events. The outputs, which we call “cues”, are combined to enable us to compute directly the probability of the object being present in the scene. An additional advantage of this approach is that the cues from different types of features are presented in a homogeneous way.
Keywords
Related Topics
Physical Sciences and Engineering
Computer Science
Signal Processing
Authors
K. Messer, W.J. Christmas, E. Jaser, J. Kittler, B. Levienaise-Obadia, D. Koubaroulis,