Abstract
This paper examines different technologies that can be applied in the design and development of a ground control station for Unmanned Aerial Vehicles (UAVs) equipped with multimodal interfaces. Multimodal technologies employ multiple sensory channels/modalities for information transmission as well as for system control. Examples of these technologies could be haptic feedback, head tracking, auditory information (3D audio), voice control, tactile displays, etc. The applicability and benefits of those technologies is analyzed for a task consisting in the acknowledgement of alerts in an UAV ground control station composed by three screens and managed by a single operator. For this purpose, several experiments were conducted with a group of individuals using different combinations of modal conditions (visual, aural and tactile).
Similar content being viewed by others
References
Biocca, F., Jin, K., Choi, Y.: Visual touch in virtual environments: an exploratory study of presence, multimodal interfaces, and cross-modal sensory illusions. Presence: Teleoperators and Virtual Environments 10(3), 247–265 (2001)
Origin Instruments Corporation: Headmouse extreme. http://www.orin.com/access/headmouse/ (2009)
Craven, P., Belov, N., Tremoulet, P., Thomas, M., Berka, C., Levendowski, D., Davis, G.: Foundations of augmented cognition, chap. cognitive workload gauge development: comparison of real-time classification methods, pp. 75–84. Springer, New York (2006)
Creative Labs: OpenAL: cross-platform 3D audio library. http://www.openal.org/ (2009)
Free Software Foundation: FreeTrack. http://www.free-track.net/english/ (2009)
Kato, T., Omachi, S., Aso, H.: Asymmetric gaussian and its application to pattern recognition. In: Proceedings of the Joint IAPR International Workshop on Structural, Syntactic, and Statistical Pattern Recognition, pp. 405–413. Springer, London (2002)
Lemon, O., Bracy, A., Gruenstein, A., Peters, S.: The WITAS multi-modal dialogue system I. In: Proceedings of the 7th European Conference on Speech Communication and Technology (EUROSPEECH), pp. 1559–1562. Aalborg, Denmark (2001)
Madentec: Tracker Pro. http://www.madentec.com/products/tracker-pro.php (2009)
Maza, I., Caballero, F.: Video summarizing the experiments reported in this paper. http://grvc.us.es/JINT_multimodal (2009)
McCarley, J.S., Wickens, C.D.: Human factors implications of UAVs in the national airspace. Tech. Rep. AHFD-05-5/FAA-05-1, Institute of Aviation, Aviation Human Factors Division, University of Illinois at Urbana-Champaign (2005)
NaturalPoint: SmartNav 4 AT. http://www.naturalpoint.com/smartnav/ (2009)
NaturalPoint: TrackIR 4. http://www.naturalpoint.com/trackir/02-products/product-TrackIR-4-PRO.html (2009)
Ollero, A., Garcia-Cerezo, A., Gomez, J.: Teleoperacion y Telerrobotica. Pearson Prentice Hall, Englewood Cliffs (2006)
Ollero, A., Maza, I. (eds.): Multiple heterogeneous unmanned aereal vehicles, chap. teleoperation tools, pp. 189–206. Springer Tracts on Advanced Robotics. Springer, New York (2007)
Orden, K.F.V., Viirre, E., Kobus, D.A.: Foundations of Augmented Cognition, chap. Augmenting Task-Centered Design with Operator State Assessment Technologies, pp. 212–219. Springer, New York (2007)
Patterson, R.D.: Guidelines for Auditory Warnings on Civil Aircraft. Civil Aviation Authority, London (1982)
Peryer, G., Noyes, J., Pleydell-Pearce, K., Lieven, N.: Auditory alert characteristics: a survey of pilot views. Int. J. Aviat. Psychol. 15(3), 233–250 (2005)
Poythress, M., Berka, C., Levendowski, D., Chang, D., Baskin, A., Champney, R., Hale, K., Milham, L., Russell, C., Seigel, S., Tremoulet, P., Craven, P.: Foundations of Augmented Cognition, chap. Correlation between expected workload and EEG indices of cognitive workload and task engagement, pp. 75–84. Springer (2006)
Sharma, R., Pavlovic, V.I., Huang, T.S.: Toward multimodal human-computer interface. Proc. IEEE 86(5), 853–869 (1998)
Cachya Software: Cachya. http://www.cachya.com/esight/overview.php (2009)
Stanford University: WITAS Project multi-modal conversational interfaces. http://www-csli.stanford.edu/semlab-hold/witas/ (2009)
Sweller, J.: Visualisation and instructional design. In: Proceedings of the International Workshop on Dynamic Visualizations and Learning (2002)
EyeTech Digital Systems: EyeTech TM3. http://www.eyetechds.com/index.htm (2009)
University of Edinburgh: The festival speech synthesis system. http://www.cstr.ed.ac.uk/projects/festival/ (2009)
Wilson, G., Russell, C.: Real-time assessment of mental workload using psychophysiological measures and artificial neural networks. In: Human Factors, pp. 635–643 (2003)
Zhu, Y.: Advances in Visual Computing, chap. Measuring Effective Data Visualization, pp. 652–661. Springer, New York (2007)
Author information
Authors and Affiliations
Corresponding author
Additional information
This work is partially supported by the CONET Network of Excellence (ICT-224053) funded by the European Commission under FP7, the ROBAIR Project funded by the Spanish Research and Development Program (DPI2008-03847) and the ATLANTIDA Project leaded by Boeing Research & Technology Europe.
Rights and permissions
About this article
Cite this article
Maza, I., Caballero, F., Molina, R. et al. Multimodal Interface Technologies for UAV Ground Control Stations . J Intell Robot Syst 57, 371–391 (2010). https://doi.org/10.1007/s10846-009-9351-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10846-009-9351-9