Learning Embeddings that Capture Spatial Semantics for Indoor Navigation - Robotics Institute Carnegie Mellon University

Learning Embeddings that Capture Spatial Semantics for Indoor Navigation

Vidhi Jain, Prakhar Agarwal, Shishir Patil, and Katia Sycara
Workshop Paper, NeurIPS '20 Workshop on Object Representations for Learning and Reasoning, December, 2020

Abstract

Incorporating domain-specific priors in search and navigation tasks has shown promising results in improving generalization and sample complexity over end-to-end trained policies.
In this work, we study how object embeddings that capture spatial semantic priors can guide search and navigation task in a structured environment.
We know that humans can search for an object like a book, or a plate in an unseen house, based on spatial semantics of bigger objects detected. For example, a book is likely to be on a bookshelf or a table, whereas a plate is likely to be in a cupboard or dishwasher. We propose a method to incorporate such spatial semantic awareness in robots by leveraging pre-trained language models and multi-relational knowledge bases as object embeddings. We demonstrate using these object embeddings to search a query object in an unseen indoor environment. We measure the performance of these embeddings in an indoor simulator (AI2Thor). We further evaluate different pre-trained embedding on Success Rate (SR) and Success weighted by Path Length (SPL).

BibTeX

@workshop{Jain-2020-129117,
author = {Vidhi Jain and Prakhar Agarwal and Shishir Patil and Katia Sycara},
title = {Learning Embeddings that Capture Spatial Semantics for Indoor Navigation},
booktitle = {Proceedings of NeurIPS '20 Workshop on Object Representations for Learning and Reasoning},
year = {2020},
month = {December},
keywords = {embeddings, navigation, search},
}