Paper
23 November 2022 Knowledge distillation for end-to-end speech recognition based on Conformer model
Jiangkun Sang, Yolwas Nurmemet
Author Affiliations +
Proceedings Volume 12454, International Symposium on Robotics, Artificial Intelligence, and Information Engineering (RAIIE 2022); 124542I (2022) https://doi.org/10.1117/12.2659711
Event: International Symposium on Robotics, Artificial Intelligence, and Information Engineering (RAIIE 2022), 2022, Hohhot, China
Abstract
With the rise of deep learning, the end-to-end speech recognition model has received increasing attention. Currently, the performance of the end-to-end speech recognition model has been further updated on basis of on the proposal of the Conformer Framework, which has been widely used in the field of speech recognition. However, these models perform poorly on edge hardware due to large memory and computation requirements. This paper compresses the model by exploring the application of knowledge distillation in end-to-end speech recognition based on Conformer. By combining with the network structure of the model, the training loss function is redefined, and on this basis, the influence of different temperatures on the performance of the student model in knowledge distillation is explored.
© (2022) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Jiangkun Sang and Yolwas Nurmemet "Knowledge distillation for end-to-end speech recognition based on Conformer model", Proc. SPIE 12454, International Symposium on Robotics, Artificial Intelligence, and Information Engineering (RAIIE 2022), 124542I (23 November 2022); https://doi.org/10.1117/12.2659711
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Performance modeling

Speech recognition

Computer programming

Convolution

Systems modeling

Transformers

Neural networks

RELATED CONTENT


Back to Top