Speech recognition has seen dramatic improvements in the last decade, though those improvements have focused primarily on adult speech. In this paper, we assess child-directed speech recognition and leverage a transfer learning approach to improve child-directed speech recognition by training the recent DeepSpeech2 model on adult data, then apply additional tuning to varied amounts of child speech data. We evaluate our model using the CMU Kids dataset as well as our own recordings of child-directed prompts. The results from our experiment show that even a small amount of child audio data improves significantly over a baseline of adult-only or child-only trained models. We report a final general Word-Error-Rate of 29% over a baseline of 62% that uses the adult-trained model. Our analyses show that our model adapts quickly using a small amount of data and that the general child model works better than school grade-specific models. We make available our trained model and our data collection tool.
Creative Commons License
This work is licensed under a Creative Commons Attribution-Noncommercial 4.0 License
Booth, Eric; Carns, Jake; Kennington, Casey; and Rafla, Nader. (2020). "Evaluating and Improving Child-Directed Automatic Speech Recognition". LREC 2020, Twelfth International Conference on Language Resources and Evaluation, 6340-6345.