Authors:
Neha Baranwal
;
Avinash Kumar Singh
and
Suna Bensch
Affiliation:
Department of Computing Science, Umeå University, Umeå and Sweden
Keyword(s):
Natural Language Grounding, Spatial Relation Extraction, Hobb’s Algorithm, Human-robot Interaction, NLTK, Google Speech, Stanford Parser.
Related
Ontology
Subjects/Areas/Topics:
Agents
;
Applications
;
Artificial Intelligence
;
Cognitive Robotics
;
Conversational Agents
;
Informatics in Control, Automation and Robotics
;
Knowledge Engineering and Ontology Development
;
Knowledge-Based Systems
;
Natural Language Processing
;
Pattern Recognition
;
Robotics and Automation
;
Symbolic Systems
Abstract:
In verbal human-robot interaction natural language utterances have to be grounded in visual scenes by the robot. Visual language grounding is a challenging task that includes identifying a primary object among several objects, together with the object properties and spatial relations among the objects. In this paper we focus on extracting this information from sentences only. We propose two language modelling techniques, one uses regular expressions and the other one utilizes Euclidian distance. We compare these two proposed techniques with two other techniques that utilize tree structures, namely an extended Hobb’s algorithm and an algorithm that utilizes a Stanford parse tree. A comparative analysis between all language modelling techniques shows that our proposed two approaches require less computational time than the tree-based approaches. All approaches perform good identifying the primary object and its property, but for spatial relation extraction the Stanford parse tree algor
ithm performs better than the other language modelling techniques. Time elapsed for the Stanford parse tree algorithm is higher than for the other techniques.
(More)