Grounding the Meaning of Words Through Vision and Interactive Gameplay

Abstract

Currently, there exists a need for simple, easily-accessible methods with which individuals lacking advanced technical training can expand and customize their robot's knowledge. This work presents a means to satisfy that need, by abstracting the task of training robots to learn about the world around them as a vision- and dialogue-based game, I Spy. In our implementation of I Spy, robots gradually learn about objects and the concepts that describe those objects through repeated gameplay. We show that I Spy is an effective approach for teaching robots how to model new concepts using representations comprised of visual attributes. The results from 255 test games show that the system was able to correctly determine which object the human had in mind 67% of the time. Furthermore, a model evaluation showed that the system correctly understood the visual representations of its learned concepts with an average of 65% accuracy. Human accuracy against the same evaluation standard was just 88% on average.

Cite

Text

Parde et al. "Grounding the Meaning of Words Through Vision and Interactive Gameplay." International Joint Conference on Artificial Intelligence, 2015.

Markdown

[Parde et al. "Grounding the Meaning of Words Through Vision and Interactive Gameplay." International Joint Conference on Artificial Intelligence, 2015.](https://mlanthology.org/ijcai/2015/parde2015ijcai-grounding/)

BibTeX

@inproceedings{parde2015ijcai-grounding,
  title     = {{Grounding the Meaning of Words Through Vision and Interactive Gameplay}},
  author    = {Parde, Natalie and Hair, Adam and Papakostas, Michalis and Tsiakas, Konstantinos and Dagioglou, Maria and Karkaletsis, Vangelis and Nielsen, Rodney D.},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2015},
  pages     = {1895-1901},
  url       = {https://mlanthology.org/ijcai/2015/parde2015ijcai-grounding/}
}