Analysing the Multiple Timescale Recurrent Neural Network for Embodied Language Understanding

Stefan Heinrich, Sven Magg, Stefan Wermter

Publikation: Konference artikel i Proceeding eller bog/rapport kapitelBidrag til bog/antologiForskningpeer review

Abstract

How the human brain understands natural language and how we can exploit this understanding for building intelligent grounded language systems is open research. Recently, researchers claimed that language is embodied in most – if not all – sensory and sensorimotor modalities and that the brain’s architecture favours the emergence of language. In this chapter we investigate the characteristics of such an architecture and propose a model based on the Multiple Timescale Recurrent Neural Network, extended by embodied visual perception, and tested in a real world scenario. We show that such an architecture can learn the meaning of utterances with respect to visual perception and that it can produce verbal utterances that correctly describe previously unknown scenes. In addition we rigorously study the timescale mechanism (also known as hysteresis) and explore the impact of the architectural connectivity in the language acquisition task
OriginalsprogEngelsk
TitelArtificial Neural Networks -- Methods and Applications in Bio-/Neuroinformatics
RedaktørerPetia D. Koprinkova-Hristova, Valeri M. Mladenov, Nikola K. Kasabov
Antal sider26
Vol/bind4
ForlagSpringer International Publishing, Switzerland
Publikationsdato1 jan. 2015
Sider149-174
DOI
StatusUdgivet - 1 jan. 2015
Udgivet eksterntJa

Fingeraftryk

Dyk ned i forskningsemnerne om 'Analysing the Multiple Timescale Recurrent Neural Network for Embodied Language Understanding'. Sammen danner de et unikt fingeraftryk.

Citationsformater