DDK : Dynamic structure pruning based on differentiable search and recursive knowledge distillation for BERT

Copyright © 2024 Elsevier Ltd. All rights reserved..

Large-scale pre-trained models, such as BERT, have demonstrated outstanding performance in Natural Language Processing (NLP). Nevertheless, the high number of parameters in these models has increased the demand for hardware storage and computational resources while posing a challenge for their practical deployment. In this article, we propose a combined method of model pruning and knowledge distillation to compress and accelerate large-scale pre-trained language models. Specifically, we introduce a dynamic structure pruning method based on differentiable search and recursive knowledge distillation to automatically prune the BERT model, named DDK. We define the search space for network pruning as all feed-forward layer channels and self-attention heads at each layer of the network, and utilize differentiable methods to determine their optimal number. Additionally, we design a recursive knowledge distillation method that employs adaptive weighting to extract the most important features from multiple intermediate layers of the teacher model and fuse them to supervise the student network learning. Our experimental results on the GLUE benchmark dataset and ablation analysis demonstrate that our proposed method outperforms other advanced methods in terms of average performance.

Medienart:

E-Artikel

Erscheinungsjahr:

2024

Erschienen:

2024

Enthalten in:

Zur Gesamtaufnahme - volume:173

Enthalten in:

Neural networks : the official journal of the International Neural Network Society - 173(2024) vom: 15. März, Seite 106164

Sprache:

Englisch

Beteiligte Personen:

Zhang, Zhou [VerfasserIn]
Lu, Yang [VerfasserIn]
Wang, Tengfei [VerfasserIn]
Wei, Xing [VerfasserIn]
Wei, Zhen [VerfasserIn]

Links:

Volltext

Themen:

Differentiable methods
Journal Article
Knowledge distillation
Model compression
Network pruning
Pre-trained models

Anmerkungen:

Date Completed 26.03.2024

Date Revised 26.03.2024

published: Print-Electronic

Citation Status MEDLINE

doi:

10.1016/j.neunet.2024.106164

funding:

Förderinstitution / Projekttitel:

PPN (Katalog-ID):

NLM368574156