Article ID Journal Published Year Pages File Type
531130 Pattern Recognition 2012 13 Pages PDF
Abstract

This paper describes a generic framework for activity recognition based on temporal signals acquired from multiple input modalities and demonstrates its use for eye–hand data fusion. As a part of the data fusion framework, we present a multi-objective Bayesian Framework for Feature Selection with a pruned-tree search algorithm for finding the optimal feature set(s) in a computationally efficient manner. Experiments on endoscopic surgical episode recognition are used to investigate the potential of using eye-tracking for pervasive monitoring of surgical operation and to demonstrate how additional information induced by hand motion can further enhance the recognition accuracy. With the proposed multi-objective BFFS algorithm, suitable feature sets both in terms of feature relevancy and redundancy can be identified with a minimal number of instruments being tracked.

► We propose a generic eye–hand fusion framework for activity recognition. ► We propose a multi-objective BFFS with pruned-tree search algorithm for finding the optimal feature set(s). ► Endoscopic surgical episode recognition experiments are performed with a combined use of eye-tracking and motion sensing. ► Optimal feature sets, in terms of feature relevancy, redundancy and number of instruments being tracked, are identified. ► We validate the framework with surgical episode recognition experiments using various types of classifiers.

Related Topics
Physical Sciences and Engineering Computer Science Computer Vision and Pattern Recognition
Authors
, , ,