کد مقاله | کد نشریه | سال انتشار | مقاله انگلیسی | نسخه تمام متن |
---|---|---|---|---|
533921 | 870190 | 2014 | 7 صفحه PDF | دانلود رایگان |
• Our detector adaptation is based on scene transformation.
• The scene transformation is robust to pitch and yaw view variations.
• Automatic preliminary camera calibration is carried out by exploiting scene information.
• Spatial–temporal voting guides the procedure of camera parameter optimization for the best performance.
This paper focuses on detecting vehicles in different target scenes with the same pre-trained detector which is very challenging due to view variations. To address this problem, we propose a novel approach for detection adaptation based on scene transformation, which contributes in both view transformation and automatic parameter estimation. Instead of modifying the pre-trained detectors, we transform scenes into frontal/rear view handling with pitch and yaw view variations. Without human interactions but only some general prior knowledge, the transformation parameters are automatically initialized, and then online optimized with spatial–temporal voting, which guarantees that the transformation matches the pre-trained detector. Since there is no need of labeling new samples and manual camera calibration, our approach can considerably reduce manual interactions. Experiments on challenging real-world videos demonstrate that our approach achieves significant improvements over the pre-trained detector, and it is even comparable to the performance of the detector trained on fully labeled sequences.
Journal: Pattern Recognition Letters - Volume 36, 15 January 2014, Pages 154–160