Multimodal Extraction and Recognition of Arabic Implicit Discourse Relations

Research output: Conference Article in Proceeding or Book/Report chapterArticle in proceedingsResearchpeer-review

Abstract

Most research on implicit discourse relation identification has focused on written language, however, it is also crucial to understand these relations in spoken discourse. We introduce a novel method for implicit discourse relation identification across both text and speech, that allows us to extract examples of semantically equivalent pairs of implicit and explicit discourse markers, based on aligning speech+transcripts with subtitles in another language variant. We apply our method to Egyptian Arabic, resulting in a novel high-quality dataset of spoken implicit discourse relations. We present a comprehensive approach to modeling implicit discourse relation classification using audio and text data with a range of different models. We find that text-based models outperform audio-based models, but combining text and audio features can lead to enhanced performance.
Original languageEnglish
Title of host publicationProceedings of the 31st International Conference on Computational Linguistics (COLING)
Number of pages4
PublisherAssociation for Computational Linguistics
Publication dateJan 2025
Pages5425-5429
Publication statusPublished - Jan 2025
EventComputational Linguistics - United Arab Emirates, Abu Dhabi, United Arab Emirates
Duration: 19 Jan 202524 Jan 2025
Conference number: 31
https://coling2025.org/

Conference

ConferenceComputational Linguistics
Number31
LocationUnited Arab Emirates
Country/TerritoryUnited Arab Emirates
CityAbu Dhabi
Period19/01/202524/01/2025
Internet address

Keywords

  • Implicit discourse relations
  • Multimodal discourse analysis
  • Spoken language processing
  • Egyptian Arabic dataset
  • Text and audio fusion

Fingerprint

Dive into the research topics of 'Multimodal Extraction and Recognition of Arabic Implicit Discourse Relations'. Together they form a unique fingerprint.

Cite this