کد مقاله | کد نشریه | سال انتشار | مقاله انگلیسی | نسخه تمام متن |
---|---|---|---|---|
531130 | 869813 | 2012 | 13 صفحه PDF | دانلود رایگان |

This paper describes a generic framework for activity recognition based on temporal signals acquired from multiple input modalities and demonstrates its use for eye–hand data fusion. As a part of the data fusion framework, we present a multi-objective Bayesian Framework for Feature Selection with a pruned-tree search algorithm for finding the optimal feature set(s) in a computationally efficient manner. Experiments on endoscopic surgical episode recognition are used to investigate the potential of using eye-tracking for pervasive monitoring of surgical operation and to demonstrate how additional information induced by hand motion can further enhance the recognition accuracy. With the proposed multi-objective BFFS algorithm, suitable feature sets both in terms of feature relevancy and redundancy can be identified with a minimal number of instruments being tracked.
► We propose a generic eye–hand fusion framework for activity recognition.
► We propose a multi-objective BFFS with pruned-tree search algorithm for finding the optimal feature set(s).
► Endoscopic surgical episode recognition experiments are performed with a combined use of eye-tracking and motion sensing.
► Optimal feature sets, in terms of feature relevancy, redundancy and number of instruments being tracked, are identified.
► We validate the framework with surgical episode recognition experiments using various types of classifiers.
Journal: Pattern Recognition - Volume 45, Issue 8, August 2012, Pages 2855–2867