Article ID Journal Published Year Pages File Type
412159 Robotics and Autonomous Systems 2015 14 Pages PDF
Abstract

•We cast the visual place recognition problem within a transfer learning framework.•A method to quantify the similarity between Source and Target data is proposed.•The Kullback–Leibler divergence and the Earth Mover’s Distance are compared.•We extend our approach to integrate informations from multiple visual cues.•We perform an extensive experimental evaluation on publicly available sequences.

One of the most impressive characteristics of human perception is its domain adaptation capability. Humans can recognize objects and places simply by transferring knowledge from their past experience. Inspired by that, current research in robotics is addressing a great challenge: building robots able to sense and interpret the surrounding world by reusing information previously collected, gathered by other robots or obtained from the web. But, how can a robot automatically understand what is useful among a large amount of information and perform knowledge transfer? In this paper we address the domain adaptation problem in the context of visual place recognition. We consider the scenario where a robot equipped with a monocular camera explores a new environment. In this situation traditional approaches based on supervised learning perform poorly, as no annotated data are provided in the new environment and the models learned from data collected in other places are inappropriate due to the large variability of visual information. To overcome these problems we introduce a novel transfer learning approach. With our algorithm the robot is given only some training data (annotated images collected in different environments by other robots) and is able to decide whether, and how much, this knowledge is useful in the current scenario. At the base of our approach there is a transfer risk measure which quantifies the similarity between the given and the new visual data. To improve the performance, we also extend our framework to take into account multiple visual cues. Our experiments on three publicly available datasets demonstrate the effectiveness of the proposed approach.

Related Topics
Physical Sciences and Engineering Computer Science Artificial Intelligence
Authors
, , , ,