TY - JOUR
T1 - Jointly improving parsing and perception for natural language commands through human-robot dialog
AU - Thomason, Jesse
AU - Padmakumar, Aishwarya
AU - Sinapov, Jivko
AU - Walker, Nick
AU - Jiang, Yuqian
AU - Yedidsion, Harel
AU - Hart, Justin
AU - Stone, Peter
AU - Mooney, Raymond J.
N1 - Funding Information:
We would like to thank our anonymous reviewers for their detailed comments and suggestions, all of which strengthened this article. This work is supported by a National Science Foundation Graduate Research Fellowship to the first author, an NSF EAGER grant (IIS-1548567), and an NSF NRI grant (IIS-1637736). This work took place in the Learning Agents Research Group (LARG) at the Artificial Intelligence Laboratory, the University of Texas at Austin. LARG research is supported in part by grants from the National Science Foundation (CNS-1305287, IIS-1637736, IIS-1651089, IIS-1724157), the Texas Department of Transportation, Intel, Raytheon, and Lockheed Martin. Peter Stone serves on the Board of Directors of Cogitai, Inc. The terms of this arrangement have been reviewed and approved by the University of Texas at Austin in accordance with its policy on objectivity in research. Portions of this research were conducted under protocols approved by the University of Texas at Austin Institutional Review Board, protocol 2013-06-0044 with PI Peter Stone and protocol 2017-01-010 with PI Justin Hart.
Publisher Copyright:
© 2020 AI Access Foundation. All rights reserved.
PY - 2020/2/1
Y1 - 2020/2/1
N2 - In this work, we present methods for using human-robot dialog to improve language understanding for a mobile robot agent. The agent parses natural language to underlying semantic meanings and uses robotic sensors to create multi-modal models of perceptual concepts like red and heavy. The agent can be used for showing navigation routes, delivering objects to people, and relocating objects from one location to another. We use dialog clarification questions both to understand commands and to generate additional parsing training data. The agent employs opportunistic active learning to select questions about how words relate to objects, improving its understanding of perceptual concepts. We evaluated this agent on Amazon Mechanical Turk. After training on data induced from conversations, the agent reduced the number of dialog questions it asked while receiving higher usability ratings. Additionally, we demonstrated the agent on a robotic platform, where it learned new perceptual concepts on the fly while completing a real-world task.
AB - In this work, we present methods for using human-robot dialog to improve language understanding for a mobile robot agent. The agent parses natural language to underlying semantic meanings and uses robotic sensors to create multi-modal models of perceptual concepts like red and heavy. The agent can be used for showing navigation routes, delivering objects to people, and relocating objects from one location to another. We use dialog clarification questions both to understand commands and to generate additional parsing training data. The agent employs opportunistic active learning to select questions about how words relate to objects, improving its understanding of perceptual concepts. We evaluated this agent on Amazon Mechanical Turk. After training on data induced from conversations, the agent reduced the number of dialog questions it asked while receiving higher usability ratings. Additionally, we demonstrated the agent on a robotic platform, where it learned new perceptual concepts on the fly while completing a real-world task.
UR - http://www.scopus.com/inward/record.url?scp=85092131934&partnerID=8YFLogxK
U2 - 10.1613/JAIR.1.11485
DO - 10.1613/JAIR.1.11485
M3 - Article
AN - SCOPUS:85092131934
VL - 67
SP - 327
EP - 374
JO - Journal of Artificial Intelligence Research
JF - Journal of Artificial Intelligence Research
SN - 1076-9757
ER -