Home
Scholarly Works
Homing in Scale Space
Conference

Homing in Scale Space

Abstract

Local visual homing is the process of determining the direction of movement required to return an agent to a goal location by comparing the current image with an image taken at the goal, known as the snapshot image. One way of accomplishing visual homing is by computing the correspondences between features and then analyzing the resulting flow field to determine the correct direction of motion. Typically, some strong assumptions need to be posited in order to compute the home direction from the flow field. For example, it is difficult to locally distinguish translation from rotation, so many authors assume rotation to be computable by other means (e.g. magnetic compass). In this paper we present a novel approach to visual homing using scale change information from Scale Invariant Feature Transforms (SIFT) which we use to compute landmark correspondences. The method described here is able to determine the direction of the goal in the robotpsilas frame of reference, irrespective of the relative 3D orientation with the goal.

Authors

Churchill D; Vardy A

Pagination

pp. 1307-1312

Publisher

Institute of Electrical and Electronics Engineers (IEEE)

Publication Date

September 1, 2008

DOI

10.1109/iros.2008.4651166

Name of conference

2008 IEEE/RSJ International Conference on Intelligent Robots and Systems

Conference proceedings

2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)

ISSN

2153-0858
View published work (Non-McMaster Users)

Contact the Experts team