Skip to main navigation Skip to search Skip to main content

Machine Reading, Fast and Slow: When Do Models 'Understand' Language?

    Research output: Conference Article in Proceeding or Book/Report chapterArticle in proceedingsResearchpeer-review

    Abstract

    Two of the most fundamental issues in Natural Language Understanding (NLU) at present are: (a) how it can established whether deep learning-based models score highly on NLU benchmarks for the 'right' reasons; and (b) what those reasons would even be. We investigate the behavior of reading comprehension models with respect to two linguistic 'skills': coreference resolution and comparison. We propose a definition for the reasoning steps expected from a system that would be 'reading slowly', and compare that with the behavior of five models of the BERT family of various sizes, observed through saliency scores and counterfactual explanations. We find that for comparison (but not coreference) the systems based on larger encoders are more likely to rely on the 'right' information, but even they struggle with generalization, suggesting that they still learn specific lexical patterns rather than the general principles of comparison.
    Original languageEnglish
    Title of host publicationProceedings of the 29th International Conference on Computational Linguistics
    Number of pages16
    Place of PublicationGyeongju, Republic of Korea
    Publication date2022
    Pages78-93
    Publication statusPublished - 2022
    EventInternational Conference on Computational Linguistics - Gyeongju, Korea, Republic of
    Duration: 12 Oct 202217 Nov 2022
    Conference number: 29th

    Conference

    ConferenceInternational Conference on Computational Linguistics
    Number29th
    Country/TerritoryKorea, Republic of
    CityGyeongju
    Period12/10/202217/11/2022

    Keywords

    • Natural Language Understanding
    • Reading Comprehension
    • Coreference Resolution
    • Comparison
    • Deep Learning Models

    Fingerprint

    Dive into the research topics of 'Machine Reading, Fast and Slow: When Do Models 'Understand' Language?'. Together they form a unique fingerprint.

    Cite this