Where Are You? Localization from Embodied Dialog

less than 1 minute read

Published:

Conference

EMNLP 2020

Authors

  • Meera Hahn
  • Jacob Krantz
  • Dhruv Batra
  • Devi Parikh
  • James M. Rehg
  • Stefan Lee
  • Peter Anderson

Contributions

  • Presents a dataset of ∼6k dialogs in which two humans with asymmetric information complete a cooperative localization task in reconstructed 3D buildings.
  • Define three challenging tasks: Localization from Embodied Dialog (LED), Embodied Visual Dialog, and Cooperative Localization.
  • Focusing on LED, we present a strong baseline model with detailed ablations characterizing both modeling choices and dataset biases.

Dataset

  • Where Are You? (WAY)
  • Built on Matterport3D

Approach

  • Localization from Embodied Dialog.
    • Dialog Representation throw LSTMs and tokenization with GloVE (standard NLP technique)
    • Environment Representation throw ResNet18 CNN.
    • Pix2Pix LingUNet
  • Embodied Visual Dialog
  • Cooperative Localization.