Embodied Language Understanding with a Multiple Timescale Recurrent Neural Network

Stefan Heinrich, Cornelius Weber, Stefan Wermter

Research output: Conference Article in Proceeding or Book/Report chapterArticle in proceedingsResearchpeer-review

Abstract

How the human brain understands natural language and what we can learn for intelligent systems is open research. Recently, researchers claimed that language is embodied in most – if not all – sensory and sensorimotor modalities and that the brain’s architecture favours the emergence of language. In this paper we investigate the characteristics of such an architecture and propose a model based on the Multiple Timescale Recurrent Neural Network, extended by embodied visual perception. We show that such an architecture can learn the meaning of utterances with respect to visual perception and that it can produce verbal utterances that correctly describe previously unknown scenes.
Original languageEnglish
Title of host publicationProceedings of the 23rd International Conference on Artificial Neural Networks (ICANN2013)
EditorsValeri Mladenov, Petia Koprinkova-Hristova, Günther Palm, Alessandro E.P. Villa, Bruno Apolloni, Nicola Kasabov
Number of pages8
Volume8131
PublisherSpringer
Publication date1 Sept 2013
Pages216-223
DOIs
Publication statusPublished - 1 Sept 2013
Externally publishedYes
SeriesLecture Notes in Computer Science

Keywords

  • Embodied Language
  • MTRNN
  • Language Acquisition

Fingerprint

Dive into the research topics of 'Embodied Language Understanding with a Multiple Timescale Recurrent Neural Network'. Together they form a unique fingerprint.

Cite this