ABSTRACT
The simple act of listening or of taking notes while attending a lesson may represent an insuperable burden for millions of people with some form of disabilities (e.g., hearing impaired, dyslexic and ESL students). In this paper, we propose an architecture that aims at automatically creating captions for video lessons by exploiting advances in speech recognition technologies. Our approach couples the usage of off-the-shelf ASR (Automatic Speech Recognition) software with a novel caption alignment mechanism that smartly introduces unique audio markups into the audio stream before giving it to the ASR and transforms the plain transcript produced by the ASR into a timecoded transcript.
- Unesco Report 2005 - The quality imperative. Global Monitoring Report, 2005. {on-line} Available at http://www.unesco.orgGoogle Scholar
- M. Xu, S. Yan, T-S. Chua, R. Hong, M. Wang. Dynamic captioning: video accessibility enhancement for hearing impairment. In Proc of the ACM Multimedia Conference, pp. 421--430, New York, NY, USA, 2010. Google ScholarDigital Library
- L. Jelinek, D. Jackson. Television literacy: comprehension of program content using closed captions for the deaf. Journal of Deaf Stud. Deaf Educ., Vol. 6, N. 1, pp. 43--53, 2001.Google Scholar
- T. Garza. Evaluating the use of captioned video materials in advanced foreign language learning. Foreign Language Annals, Vol. 24, N. 3, pp. 239--258, May 1991.Google ScholarCross Ref
- S. Tsuboi, N. Shimogori, T. Ikeda. Automatically generated captions: will they help non-native speakers communicate in english? In Proc of Intercultural collaboration conference, ICIC '10, pp. 79--86, New York, NY, USA, 2010. ACM. Google ScholarDigital Library
- G. Penn, E. Toms, D. James, C. Munteanu, R. Baecker. The effect of speech recognition accuracy rates on the usefulness and usability of webcast archives. In Proc of the SIGCHI conference on Human Factors in computing systems, CHI '06, pp. 493--502, New York, NY, USA, 2006. ACM. Google ScholarDigital Library
- M. Wald. Crowdsourcing correction of speech recognition captioning errors. In Proc of the W4A Conference, New York, NY, USA, 2011. ACM. Google ScholarDigital Library
- A. Knight, K. C. Almeroth. Fast caption alignment for automatic indexing of audio. International Journal of Multimedia Data Engineering and Management, Vol. 1, N. 2, pp. 1--17. June 2010. Google ScholarDigital Library
Index Terms
- Enhancing learning accessibility through fully automatic captioning
Recommendations
Dynamic captioning: video accessibility enhancement for hearing impairment
MM '10: Proceedings of the 18th ACM international conference on MultimediaThere are more than 66 million people su®ering from hearing impairment and this disability brings them di±culty in the video content understanding due to the loss of audio information. If scripts are available, captioning technology can help them in a ...
Enhancing the accessibility of e-learning platforms through synthetic speech
PCI '16: Proceedings of the 20th Pan-Hellenic Conference on InformaticsThis paper describes the design and development considerations for the adaptation of an accessibility enhancement tool to the Open eClass e-learning platform. The tool aims to improve the accessibility of the platform for the visually challenged ...
Universal access to communication and learning: the role of automatic speech recognition
This communication discusses how automatic speech recognition (ASR) can support universal access to communication and learning through the cost-effective production of text synchronised with speech and describes achievements and planned developments of ...
Comments