Abstract

This paper describes a probabilistic framework for appearance based navigation and mapping using spatial and visual appearance data. Like much recent work on appearance based navigation we adopt a bag-of-words approach in which positive or negative observations of visual words in a scene are used to discriminate between already visited and new places. In this paper we add an important extra dimension to the approach. We explicitly model the spatial distribution of visual words as a random graph in which nodes are visual words and edges are distributions over distances. Care is taken to ensure that the spatial model is able to capture the multi-modal distributions of inter-word spacing and account for sensor errors both in word detection and distances. Crucially, these inter-word distances are viewpoint invariant and collectively constitute strong place signatures and hence the impact of using both spatial and visual appearance is marked. We provide results illustrating a tremendous increase in precision-recall area compared to a state-of-the-art visual appearance only systems.

Keywords

Computer scienceArtificial intelligenceProbabilistic logicComputer visionInvariant (physics)Precision and recallPattern recognition (psychology)Word (group theory)VisualizationDimension (graph theory)MathematicsGeometry

Affiliated Institutions

Related Publications

Publication Info

Year
2010
Type
article
Pages
2649-2656
Citations
122
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

122
OpenAlex

Cite This

Rohan Paul, Paul Newman (2010). FAB-MAP 3D: Topological mapping with spatial and visual appearance. , 2649-2656. https://doi.org/10.1109/robot.2010.5509587

Identifiers

DOI
10.1109/robot.2010.5509587