Abstract
Traditional language model compression techniques, like knowledge distillation, require a fixed architecture, limiting flexibility, while structured pruning methods often fail to preserve performance. This paper introduces Iterative Structured Knowledge Distillation (ISKD), which integrates knowledge distillation and structured pruning by progressively replacing transformer blocks with smaller, efficient versions during training. This study validates ISKD on two transformer-based language models: GPT-2 and Phi-1. ISKD outperforms L1 pruning and achieves similar performance to knowledge distillation while offering greater flexibility. ISKD reduces model parameters - 30.68% for GPT-2 and 30.16% for Phi-1 - while maintaining at least four-fifths of performance on both language modeling and commonsense reasoning tasks. These findings suggest that this method offers a promising balance between model efficiency and accuracy.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 31st International Conference on Computational Linguistics |
| Editors | Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert |
| Number of pages | 6 |
| Place of Publication | Abu Dhabi, UAE |
| Publisher | Association for Computational Linguistics |
| Publication date | 1 Jan 2025 |
| Pages | 6601-6606 |
| Publication status | Published - 1 Jan 2025 |
| Event | International Conference on Computational Linguistics - Abu Dhabi, United Arab Emirates Duration: 19 Jan 2025 → 24 Jan 2025 Conference number: 31 https://coling2025.org/ https://coling2025.org/calls/main_conference_papers/ |
Conference
| Conference | International Conference on Computational Linguistics |
|---|---|
| Number | 31 |
| Country/Territory | United Arab Emirates |
| City | Abu Dhabi |
| Period | 19/01/2025 → 24/01/2025 |
| Internet address |
Keywords
- Iterative Structured Knowledge Distillation
- Structured pruning
- Transformer block replacement
- Language model compression
- Parameter-efficient NLP
Fingerprint
Dive into the research topics of 'Iterative Structured Knowledge Distillation: Optimizing Language Models Through Layer-by-Layer Distillation'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver