Abstract
Patients who have undergone total laryngectomy and use electrolarynx for voice production suffer from poor intelligibility. It may lead in many cases to fear of speaking to strangers, even over the phone. Automatic Speech Recognition (ASR) systems could help patients overcome this problem in many ways. Unfortunately, even state-of-the-art ASR systems cannot provide results comparable to those of conventional speakers. The problem is mainly caused by the similarity between voiced and unvoiced phoneme pairs. In many cases, a language model can help to solve the issue, but only if the word context is sufficiently long. Therefore adjustment of acoustic data and/or acoustic model is necessary to increase recognition accuracy. In this paper, we propose voiceless phonemes elongation to improve recognition accuracy and enrich the ASR system with a model that takes this elongation into account. The idea of elongation is verified on a set of ASR experiments with artificially elongated voiceless phonemes. To enriching the ASR system, the DNN model for rescoring lattices based on phoneme duration is proposed. The new system is compared with a standard ASR. It is also verified that the ASR system created using elongated synthetic data can successfully recognize the actual elongated data pronounced by the real speaker.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
It is essential to mention that all speech data are in the Czech language.
- 2.
scythe.
- 3.
goat.
- 4.
The duration model trained and tested only on artificial data achieved \(Acc_{p} = 88.54\%\).
References
Denby, B., Schultz, T., Honda, K., Hueber, T., Gilbert, J.M., Brumberg, J.: Silent speech interfaces. Speech Commun. 52(4), 270–287 (2010). https://doi.org/10.1016/j.specom.2009.08.002, http://linkinghub.elsevier.com/retrieve/pii/S0167639309001307
Hadian, H., Povey, D., Sameti, H., Khudanpur, S.: Phone duration modeling for LVCSR using neural networks. In: INTERSPEECH, pp. 518–522 (2017)
Liu, H., Ng, M.L.: Electrolarynx in voice rehabilitation. Auris Nasus Larynx 34(3), 327–332 (2007). https://doi.org/10.1016/j.anl.2006.11.010
Radová, V., Psutka, J.: UWB-S01 corpus: a czech read-speech corpus. In: Proceedings of the 6th International Conference on Spoken Language Processing, pp. 732–735. ICSLP2000 (2000)
Stanislav, P., Psutka, J.V.: Influence of different phoneme mappings on the recognition accuracy of electrolaryngeal speech. In: Proceedings of the International Conference on Signal Processing and Multimedia Applications and Wireless Information Networks and Systems - Volume 1: SIGMAP, (ICETE 2012), pp. 204–207 (2012). https://doi.org/10.5220/0004129502040207
Stanislav, P., Psutka, J.V., Psutka, J.: Recognition of the electrolaryngeal speech: comparison between human and machine. In: Ekštein, K., Matoušek, V. (eds.) TSD 2017. LNCS (LNAI), vol. 10415, pp. 509–517. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-64206-2_57
Acknowledgments
This research was supported by the Technology Agency of the Czech Republic, project No. TN01000024 and by the grant of the University of West Bohemia, project No. SGS-2019-027.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Stanislav, P., Psutka, J.V., Psutka, J. (2020). Increasing the Accuracy of the ASR System by Prolonging Voiceless Phonemes in the Speech of Patients Using the Electrolarynx. In: Karpov, A., Potapova, R. (eds) Speech and Computer. SPECOM 2020. Lecture Notes in Computer Science(), vol 12335. Springer, Cham. https://doi.org/10.1007/978-3-030-60276-5_54
Download citation
DOI: https://doi.org/10.1007/978-3-030-60276-5_54
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-60275-8
Online ISBN: 978-3-030-60276-5
eBook Packages: Computer ScienceComputer Science (R0)