JOURNAL ARTICLE

Learning word pronunciations using a recurrent neural network

Abstract

Segmentation is the process of dividing a printed character string into graphemes, each of which is associated with one (or rarely more) output phonemes. The purpose of this study was to investigate what internal representation of the segmentation process and character-to-phoneme correspondences would be learned by a recurrent neural network as it was trained to produce the correct temporal sequence of phonemes for printed words held fixed on its input nodes. The resilient recurrent backpropagation network learned very effectively to generate the correct pronunciation for 150 words. Some interesting rules of pronunciation discovered by the network were extracted despite the network's distributed representation.

Keywords:
Pronunciation Computer science Character (mathematics) Backpropagation Artificial intelligence Recurrent neural network Artificial neural network Natural language processing String (physics) Speech recognition Word (group theory) Representation (politics) Segmentation Process (computing) Text segmentation Sequence (biology) Time delay neural network Pattern recognition (psychology) Linguistics

Metrics

6
Cited By
0.37
FWCI (Field Weighted Citation Impact)
10
Refs
0.62
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Neural Networks and Applications
Physical Sciences →  Computer Science →  Artificial Intelligence
Speech Recognition and Synthesis
Physical Sciences →  Computer Science →  Artificial Intelligence
Handwritten Text Recognition Techniques
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
© 2026 ScienceGate Book Chapters — All rights reserved.