Skip to main navigation Skip to search Skip to main content

IndicBART: A Pre-trained Model for Indic Natural Language Generation.

  • Raj Dabre
  • , Himani Shrotriya
  • , Anoop Kunchukuttan
  • , Ratish Puduppully
  • , Mitesh M. Khapra
  • , Pratyush Kumar

Research output: Conference Article in Proceeding or Book/Report chapterArticle in proceedingsResearchpeer-review

Abstract

In this paper, we study pre-trained sequence-to-sequence models for a group of related languages, with a focus on Indic languages. We present IndicBART, a multilingual, sequence-to-sequence pre-trained model focusing on 11 Indic languages and English. IndicBART utilizes the orthographic similarity between Indic scripts to improve transfer learning between similar Indic languages. We evaluate IndicBART on two NLG tasks: Neural Machine Translation (NMT) and extreme summarization. Our experiments on NMT and extreme summarization show that a model specific to related languages like IndicBART is competitive with large pre-trained models like mBART50 despite being significantly smaller. It also performs well on very low-resource translation scenarios where languages are not included in pre-training or fine-tuning. Script sharing, multilingual training, and better utilization of limited model capacity contribute to the good performance of the compact IndicBART model.
Original languageEnglish
Title of host publicationFindings of the Association for Computational Linguistics: ACL 2022
PublisherAssociation for Computational Linguistics
Publication date2022
Pages1849-1863
DOIs
Publication statusPublished - 2022
Externally publishedYes
EventConference on the Association for Computational Linguistics - Dublin, Ireland
Duration: 22 May 202227 May 2022
Conference number: 60

Conference

ConferenceConference on the Association for Computational Linguistics
Number60
Country/TerritoryIreland
CityDublin
Period22/05/202227/05/2022

Keywords

  • IndicBART
  • Multilingual sequence-to-sequence models
  • Orthographic similarity
  • Neural machine translation
  • Extreme summarization

Fingerprint

Dive into the research topics of 'IndicBART: A Pre-trained Model for Indic Natural Language Generation.'. Together they form a unique fingerprint.

Cite this