back icon
close icon

Capture phrases in quotes for more specific queries (e.g. "rocket ship" or "Fred Lynn")

Conference Paper  September 9, 2019

Semantically-Aware Attentive Neural Embeddings for 2D Long-Term Visual Localization

SRI Authors Han-Pang Chiu, Supun Samarasekera, Rakesh “Teddy” Kumar

Citation

COPY

Zachary Seymour, Karan Sikka, Han-Pang Chiu, Supun Samarasekera, Rakesh Kumar: Semantically-Aware Attentive Neural Embeddings for 2D Long-Term Visual Localization. BMVC 2019: 70

Abstract

We present an approach that combines appearance and semantic information for 2D image-based localization (2D-VL) across large perceptual changes and time lags. Compared to appearance features, the semantic layout of a scene is generally more invariant to appearance variations. We use this intuition and propose a novel end-to-end deep attention-based framework that utilizes multimodal cues to generate robust embeddings for 2D-VL. The proposed attention module predicts a shared channel attention and modality-specific spatial attentions to guide the embeddings to focus on more reliable image regions. We evaluate our model against state-of-the-art (SOTA) methods on three challenging localization datasets. We report an average (absolute) improvement of 19% over current SOTA for 2D-VL. Furthermore, we present an extensive study demonstrating the contribution of each component of our model, showing 8–15% and 4% improvement from adding semantic information and our proposed attention module. We finally show the predicted attention maps to offer useful insights into our model.  

How can we help?

    Once you hit send…

    We’ll match your inquiry to the person who can best help you. Expect a response within 48 hours.

    Our Privacy Policy