A speech recognition method with enhanced transformer decoder
Abstract Addressing the issue that the Transformer decoder struggles to capture local features for monotonic alignment in speech recognition, and simultaneously incorporating language model cold fusion training into the decoder, an enhanced decoder-based speech recognition model is investigated. The...
Saved in:
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
SpringerOpen
2025-02-01
|
Series: | EURASIP Journal on Audio, Speech, and Music Processing |
Subjects: | |
Online Access: | https://doi.org/10.1186/s13636-025-00394-6 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1823861721475842048 |
---|---|
author | Hengbo Hu Tong Niu Zhenhua He |
author_facet | Hengbo Hu Tong Niu Zhenhua He |
author_sort | Hengbo Hu |
collection | DOAJ |
description | Abstract Addressing the issue that the Transformer decoder struggles to capture local features for monotonic alignment in speech recognition, and simultaneously incorporating language model cold fusion training into the decoder, an enhanced decoder-based speech recognition model is investigated. The enhanced decoder separates and combines the two attention mechanisms in the Transformer decoder into cross-attention layers and a self-attention language model module. The cross-attention layers are utilized to capture local features more efficiently from the encoder output, and the self-attention language model module is used to pre-train with additional domain-related text, followed by cold fusion training. Experimental results on the Mandarin Aishell-1 dataset demonstrate that when the encoder is a Conformer, the enhanced decoder achieves a 16.1% reduction in character error rate compared to the Transformer decoder. Furthermore, when the language model is pre-trained with suitable text data, the performance of the cold fusion-trained model is further enhanced. |
format | Article |
id | doaj-art-5b6fd2d632584e43854bd517d0fc49eb |
institution | Kabale University |
issn | 1687-4722 |
language | English |
publishDate | 2025-02-01 |
publisher | SpringerOpen |
record_format | Article |
series | EURASIP Journal on Audio, Speech, and Music Processing |
spelling | doaj-art-5b6fd2d632584e43854bd517d0fc49eb2025-02-09T12:48:48ZengSpringerOpenEURASIP Journal on Audio, Speech, and Music Processing1687-47222025-02-012025111210.1186/s13636-025-00394-6A speech recognition method with enhanced transformer decoderHengbo Hu0Tong Niu1Zhenhua He2Research and Development Department 1 - Intelligent Speech Technology Team, Zhengzhou Xinda Institute of Advanced TechnologySchool of Information Systems Engineering, University of Information EngineeringResearch and Development Department 1 - Intelligent Speech Technology Team, Zhengzhou Xinda Institute of Advanced TechnologyAbstract Addressing the issue that the Transformer decoder struggles to capture local features for monotonic alignment in speech recognition, and simultaneously incorporating language model cold fusion training into the decoder, an enhanced decoder-based speech recognition model is investigated. The enhanced decoder separates and combines the two attention mechanisms in the Transformer decoder into cross-attention layers and a self-attention language model module. The cross-attention layers are utilized to capture local features more efficiently from the encoder output, and the self-attention language model module is used to pre-train with additional domain-related text, followed by cold fusion training. Experimental results on the Mandarin Aishell-1 dataset demonstrate that when the encoder is a Conformer, the enhanced decoder achieves a 16.1% reduction in character error rate compared to the Transformer decoder. Furthermore, when the language model is pre-trained with suitable text data, the performance of the cold fusion-trained model is further enhanced.https://doi.org/10.1186/s13636-025-00394-6Cross-attentionTransformer decoderLanguage model cold fusion |
spellingShingle | Hengbo Hu Tong Niu Zhenhua He A speech recognition method with enhanced transformer decoder EURASIP Journal on Audio, Speech, and Music Processing Cross-attention Transformer decoder Language model cold fusion |
title | A speech recognition method with enhanced transformer decoder |
title_full | A speech recognition method with enhanced transformer decoder |
title_fullStr | A speech recognition method with enhanced transformer decoder |
title_full_unstemmed | A speech recognition method with enhanced transformer decoder |
title_short | A speech recognition method with enhanced transformer decoder |
title_sort | speech recognition method with enhanced transformer decoder |
topic | Cross-attention Transformer decoder Language model cold fusion |
url | https://doi.org/10.1186/s13636-025-00394-6 |
work_keys_str_mv | AT hengbohu aspeechrecognitionmethodwithenhancedtransformerdecoder AT tongniu aspeechrecognitionmethodwithenhancedtransformerdecoder AT zhenhuahe aspeechrecognitionmethodwithenhancedtransformerdecoder AT hengbohu speechrecognitionmethodwithenhancedtransformerdecoder AT tongniu speechrecognitionmethodwithenhancedtransformerdecoder AT zhenhuahe speechrecognitionmethodwithenhancedtransformerdecoder |