JOURNAL ARTICLE

Sequence Noise Injected Training for End-to-end Speech Recognition

Abstract

We present a simple noise injection algorithm for training end-to-end ASR models which consists in adding to the spectra of training utterances the scaled spectra of random utterances of comparable length. We conjecture that the sequence information of the "noise" utterances is important and verify this via a contrast experiment where the frames of the utterances to be added are randomly shuffled. Experiments for both CTC and attention-based models show that the pro-posed scheme results in up to 9% relative word error rate improvements (depending on the model and test set) on the Switchboard 300 hours English conversational telephony database. Additionally, we set a new benchmark for attention-based encoder-decoder models on this corpus.

Keywords:
Computer science Speech recognition Noise (video) Telephony Encoder Benchmark (surveying) Sequence (biology) Set (abstract data type) Word error rate Word (group theory) Training set Test set End-to-end principle Decoding methods Artificial intelligence Encoding (memory) Pattern recognition (psychology) Algorithm Mathematics Telecommunications

Metrics

43
Cited By
4.30
FWCI (Field Weighted Citation Impact)
40
Refs
0.95
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Speech Recognition and Synthesis
Physical Sciences →  Computer Science →  Artificial Intelligence
Speech and Audio Processing
Physical Sciences →  Computer Science →  Signal Processing
Music and Audio Processing
Physical Sciences →  Computer Science →  Signal Processing
© 2026 ScienceGate Book Chapters — All rights reserved.