Question Answering with Hierarchical Attention Networks

Tayfun Alpay, Stefan Heinrich, Michael Nelskamp, Stefan Wermter

Research output: Conference Article in Proceeding or Book/Report chapterArticle in proceedingsResearchpeer-review

Abstract

We investigate hierarchical attention networks for the task of question answering. For this purpose, we propose two different approaches: in the first, a document vector representation is built hierarchically from word-to-sentence level which is then used to infer the right answer. In the second, pointer sum attention is utilized to directly infer an answer from the attention values of the word and sentence representations. We evaluate our approach on the Children's Book Test, a cloze-style question answering dataset, and analyze the generated attention distributions. Our results show that, although a hierarchical approach does not offer much improvement over a shallow baseline, it does indeed offer a large performance boost when combining word and sentence attention with pointer sum attention.
Original languageEnglish
Title of host publicationProceedings of the International Joint Conference on Neural Networks (IJCNN2019)
Number of pages6
Place of PublicationBudapest, Hungary
Publication date1 Sept 2019
Pages13-18
DOIs
Publication statusPublished - 1 Sept 2019
Externally publishedYes

Keywords

  • Knowledge discovery
  • Task analysis
  • Encoding
  • Adaptation models
  • Computer architecture
  • Neural networks
  • Training

Fingerprint

Dive into the research topics of 'Question Answering with Hierarchical Attention Networks'. Together they form a unique fingerprint.

Cite this