Abstract
In this paper, we present a 3D User Interface design for a shared augmented reality setup to support strategic, tactical, and training platforms. An accurate 3D holographic object registration is necessary to enable a shared augmented reality experience using multiple Microsoft HoloLens devices. We developed a sensor data fusion framework which uses both external positional sensor data and Microsoft HoloLens to reduce augmented reality registration errors in our shared augmented reality application. Our shared augmented reality 3D User Interface design implemented in the application, together with our sensor data fusion framework, enables the use of multiple Microsoft HoloLens to support a mission planning scenario.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Su, S., et al.: Sensor data fusion framework to improve holographic object registration accuracy for a shared augmented reality mission planning scenario. In: Chen, J.Y., Frago-meni, G. (eds.) Virtual. Augmented and Mixed Reality: Interaction, Navigation, Visualization, Embodiment, and Simulation, pp. 202–214. Springer International Publishing, Cham (2018)
Stasser, G., Titus, W.: Pooling of unshared information in group decision making: biased information sampling during discussion. J. Pers. Soc. Psychol. 48(6), 1467–1478 (1985). https://doi.org/10.1037/0022-3514.48.6.1467
Kiesler, S., Siegel, J., McGuire, T.W.: Social psychological aspects of computer-mediated communication. Am. Psychol. 39(10), 1123–1134 (1984). https://doi.org/10.1037/0003-066X.39.10.1123
Hosobori, A., Kakehi, Y.: Eyefeel & Eyechime: a face to face communication environment by augmenting eye gaze information. In: Proceedings of the 5th Augmented Human International Conference, AH 2014, pp. 7:1–7:4. ACM, New York (2014). https://doi.org/10.1145/2582051.2582058
Furlan, R.: The future of augmented reality: Hololens - Microsoft’s AR headset shines despite rough edges. IEEE Spectr. 53(6), 21 (2016). https://doi.org/10.1109/MSPEC.2016.7473143
Leicht, R.M., Maldovan, K., Messner, J.I.: A framework to analyze the effectiveness of team interactions in virtual environments. In: Proceedings of the 7th International Conference on Construction Applications of Virtual Reality. University Park, PA, USA, 2007
Shachaf, P., Hara, N.: Team effectiveness in virtual environments: an ecological approach. In: Teaching and Learning with Virtual Teams, pp. 83–108. Idea Group Publishing (2005)
Hasler, B.S., Buecheler, T., Pfeifer, R.: Collaborative work in 3D virtual environments: a research agenda and operational framework. In: Ozok, A.A., Zaphiris, P. (eds.) Online Communities and Social Computing, pp. 23–32. Springer, Heidelberg (2009)
Wagner, M.T., Regenbrecht, H.T.: Shared reality meeting - a collaborative augmented reality environment. In: The First IEEE International Workshop Augmented Reality Toolkit, p. 2, September 2002. https://doi.org/10.1109/art.2002.1106970
Regenbrecht, H.T., Wagner, M.T.: Interaction in a collaborative augmented reality environment. In: CHI 2002 Extended Abstracts on Human Factors in Computing Systems, CHI EA 2002, pp. 504–505. ACM, New York (2002). https://doi.org/10.1145/506443.506451
Billinghurst, M., et al.: Mixing realities in shared space: an augmented reality interface for collaborative computing. In: 2000 IEEE International Conference on Multimedia and Expo, ICME 2000. Proceedings. Latest Advances in the Fast Changing Worldof Multimedia (Cat. No. 00TH8532), vol. 3, pp. 1641–1644, July 2000
Piumsomboon, T., Dey, A., Ens, B., Lee, G., Billinghurst, M.: [POSTER] CoVAR: mixed-platform remote collaborative augmented and virtual realities system with shared collaboration cues. In: 2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct), pp. 218–219, October 2017. https://doi.org/10.1109/ismar-adjunct.2017.72
Oda, O., Feiner, S.: 3D referencing techniques for physical objects in shared augmented reality. In 2012 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), pp. 207–215, Nov 2012. https://doi.org/10.1109/ismar.2012.6402558
Su, S., et al.: Reconfigurable visual computing architecture for extreme-scale visual analytics. In: Proceedings of the SPIE 10652, Disruptive Technologies in Information Sciences, 106520M, 9 May 2018. https://doi.org/10.1117/12.2303887
Acknowledgments
This work was supported in part by the DOD High Performance Computing Modernization Program at The Army Research Laboratory (ARL), Department of Defense Supercomputing Resource Center (DSRC).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 This is a U.S. government work and not under copyright protection in the U.S.; foreign copyright protection may apply
About this paper
Cite this paper
Su, S., Perry, V., Roy, H., Gamble, K., Kase, S. (2020). 3D User Interface for a Multi-user Augmented Reality Mission Planning Application. In: Cassenti, D. (eds) Advances in Human Factors and Simulation. AHFE 2019. Advances in Intelligent Systems and Computing, vol 958. Springer, Cham. https://doi.org/10.1007/978-3-030-20148-7_12
Download citation
DOI: https://doi.org/10.1007/978-3-030-20148-7_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-20147-0
Online ISBN: 978-3-030-20148-7
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)