Abstract
The project and the first prototype of an interface for dictating, recording and printing radiological reports is presented. The most important feature of this interface is multimodality. The radiologist may choose among speech, keyboard and mouse to generate a report. If he is busy with hands and eyes, he can dictate most of the report, when holding and analyzing the radiographs. He can type the text and select some actions from a menu or click some meaningful icons when he has his hands free. The choice of the input modality depends on the “freedom” of the radiologist and the easiness and quickness of communication. The motivation for such a project, and the study of the impact of this system on the organisation of the radiologic department in terms of possible improvements on the reporting service, are also presented. The constraints on the radiologist-computer speech communication are analyzed. The interface has been tested with the speech modality. The speech recognizer has been trained on two main dictionaries obtained by processing chest and US reports. Four users pronounced 100 chest and 20 US reports, that have been used as a speech data-test for the automatic speech recognizer. The recognition rate of the speech recognizer, on the two dictionaries, has been 98% in the best case (speaker SP4) and 93% in the worst case (speaker SP1).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
G. Lazzari, “Riconoscimento del parlato: nuovi problemi e prospettive di ricerca”,Sistemi di Telecomunicazione, June 1989.
A. Robbins, D. Horowitz, “Speech-Controlled Generation of Radiology Reports” ¡Radiology, 164, n. 2, pp. 569–573.
Medical Applications of Voice Responsfe Technology Proceedings, Pittsburgh (PA), December 5–6, 1989.
G. Dunham, “The Role of Syntax in the Sublanguage of Medical Diagnostic Statements”, Analyzing Language in Restricted Domains, LEA, London, 1986.
L. Hirschman, Discovering Sublanguage Structures, Analyzing Language in Restricted Domains, LEA, London, 1986.
G. Antoniol, F. Dalla Palma, G. Lazzari, E. Moser “Un Sistema per la Dettatura Automatica di Referti Radiologici”, Proceedings of AIIM 90, Ed. Franco Angeli, pp. 21–28.
P. Alto, M. Brandetti,M. Ferretti, G. Maltese, F. Mancini, A. Mazza, S. Scarci, G. Vitiliano, “Adapting a Large Vocabulary Speech Recognition System to Different Tasks”, Proceedings of EUSIPCO 90, pp. 1379–1382.
L.R. Bahl, F. Jelinek, R.L. Mercer, “A Maximum Likelihood Approach to Continuous Speech Recognition”, IEEE Transactions on PAMI, vol. 5, no. 2, 1983, pp. 179–190.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1991 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Antoniol, G., Brugnara, F., Dalla Palma, F., Lazzari, G., Moser, E. (1991). A.RE.S.:An Interface for Automatic Reporting by Speech. In: Adlassnig, KP., Grabner, G., Bengtsson, S., Hansen, R. (eds) Medical Informatics Europe 1991. Lecture Notes in Medical Informatics, vol 45. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-93503-9_26
Download citation
DOI: https://doi.org/10.1007/978-3-642-93503-9_26
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-54392-3
Online ISBN: 978-3-642-93503-9
eBook Packages: Springer Book Archive