Journal of Jilin University Science Edition ›› 2024, Vol. 62 ›› Issue (1): 122-0131.

Previous Articles     Next Articles

Compression Algorithms for Automatic Speech Recognition Models: A Survey

SHI Xiaohu1, YUAN Yuping2, LV Guilin3, CHANG Zhiyong4, ZOU Yuanjun5   

  1. 1. College of Computer Science and Technology, Jilin University, Changchun 130012, China; 2. Management Center of Big Data and Network, Jilin University, Changchun 130012, China; 3. Intelligent Network Development Institute, R&D Institute of China FAW Group Co., Ltd, Changchun 130011, China; 4. College of Biological and Agricultural Engineering, Jilin University, Changchun 130022, China;
    5. School of Medical Information, Changchun University of Chinese Medicine, Changchun 130117, China
  • Received:2023-02-23 Online:2024-01-26 Published:2024-01-26

Abstract: With the development of deep learning technology,  the number of parameters in automatic speech recognition task  models was becoming  increasingly  large, which gradually increased  the computing overhead, storage requirements and power consumption of the models, and it was difficult to deploy on resource-constrained devices. Therefore, it was of great  value to compress the automatic speech recognition models based on deep learning to reduce the size of the modes while maintaining the original performance as much as possible. Aiming at the above problems,  a comprehensive survey was conducted on  the main works in this field in recent years, which was summarized as several methods, including knowledge distillation, model quantization, low-rank decomposition, network pruning, parameter sharing and combination models, and  conducted a systematic review  to  provide alternative solutions for the deployment of models on resource-constrained devices.

Key words: speech recognition, model compression, knowledge distillation, model quantization, low-rank decomposition, network pruning, parameter sharing

CLC Number: 

  • TP391