Iterative Structured Knowledge Distillation: Optimizing Language Models Through Layer-by-Layer Distillation

Malthe Have Musaeus, Rob van der Goot

Publikation: Konference artikel i Proceeding eller bog/rapport kapitelKonferencebidrag i proceedingsForskningpeer review

Abstract

Traditional language model compression techniques, like knowledge distillation, require a fixed architecture, limiting flexibility, while structured pruning methods often fail to preserve performance. This paper introduces Iterative Structured Knowledge Distillation (ISKD), which integrates knowledge distillation and structured pruning by progressively replacing transformer blocks with smaller, efficient versions during training. This study validates ISKD on two transformer-based language models: GPT-2 and Phi-1. ISKD outperforms L1 pruning and achieves similar performance to knowledge distillation while offering greater flexibility. ISKD reduces model parameters - 30.68% for GPT-2 and 30.16% for Phi-1 - while maintaining at least four-fifths of performance on both language modeling and commonsense reasoning tasks. These findings suggest that this method offers a promising balance between model efficiency and accuracy.
OriginalsprogEngelsk
TitelProceedings of the 31st International Conference on Computational Linguistics
RedaktørerOwen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Antal sider6
UdgivelsesstedAbu Dhabi, UAE
ForlagAssociation for Computational Linguistics
Publikationsdato1 jan. 2025
Sider6601-6606
StatusUdgivet - 1 jan. 2025

Fingeraftryk

Dyk ned i forskningsemnerne om 'Iterative Structured Knowledge Distillation: Optimizing Language Models Through Layer-by-Layer Distillation'. Sammen danner de et unikt fingeraftryk.

Citationsformater