Skip to main content

An Efficient Geometric Approach for Occlusion Handling in Outdoors Augmented Reality Applications

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 9768))

Abstract

Mobile location-based AR frameworks typically project information about real or virtual locations in the vicinity of the user. Those locations are treated indiscriminately, regardless of whether they are actually within the field of view (FoV) of the user or not. However, displaying occluded objects often misleads users’ perception thereby compromising the clarity and explicitness of AR applications. This paper introduces an efficient geometric technique aiming at assisting developers of outdoors mobile AR applications in generating a realistic FoV for the users. Our technique enables real time building recognition in order to address the occlusion of physical or virtual objects by physical artifacts. Our method is demonstrated in the location-based AR game Order Elimination. The latter utilizes publicly available building information to calculate the players’ FoV in real-time. Extensive performance tests provide sufficient evidence that real-time FoV rendering is feasible by modest mobile devices, even under stress operation conditions. A user evaluation study reveals that the consideration of buildings for determining FoV in mobile AR games can increase the quality of experience perceived by players when compared with standard FoV generation methods.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    The term ‘first-person shooter’ refers to a video game genre centered on gun and projectile weapon-based combat through a first-person perspective; that is, the player experiences the action through the eyes of the protagonist.

  2. 2.

    https://store.arduino.cc/product/SEN136B5B.

  3. 3.

    https://www.google.com/atap/project-tango/.

  4. 4.

    http://structure.io.

  5. 5.

    A commercial version of Order Elimination is available in Google Play (https://play.google.com/store/apps/details?id=bl.on.mi.en).

  6. 6.

    http://overpass-turbo.eu/.

  7. 7.

    The OSM Overpass API provides textual description of buildings in a certain area, essentially a list of polygons, each comprising a series of latitude/longitude points.

  8. 8.

    http://www.beyondar.com/.

  9. 9.

    Bearing refers to the angle of a moving object’s direction from the North.

  10. 10.

    As reported in the TimeWarp evaluation results, the virtual artifacts in AR games should be preferably hidden when the player has no LoS with them. Even though we could effortlessly hide the AR content when eye contact with the Zombie was infeasible, we chose not to for game design purposes.

  11. 11.

    Vuzix M100 Smart Glasses (https://www.vuzix.com/Products/M100-Smart-Glasses) supports a FoV angle of 15°, while Epson Moverio BT-200 (http://global.epson.com/newsroom/2014/news_20140107.html) supports 28°. Both products fully support smartphone integration.

  12. 12.

    The Android Augment Reality Framework (https://github.com/phishman3579/android-augment-reality-framework) and Mixare (http://www.mixare.org/) support a FoV angle of 45°.

  13. 13.

    The intuition for choosing the 5° angle has been, firstly, to create a sufficient FoV polygon for the Zombie to be included in and, secondly, to allow sufficiently wide FoV (than that of a single ray) so as to compensate for GPS location fix inaccuracies, which do not allow precise positioning calculations as in computer games.

  14. 14.

    The participants have been recruited through an open invitation advertised in the University of the Aegean, Mytilene, Greece. Ten (10) of the participants have been male and two (2) female. Six (6) participants have been in the age group of 19–23, four (4) in the age group of 24–33 and two (2) in the age group 34–39.

  15. 15.

    A video of the application can be found at https://www.youtube.com/watch?v=D–A3fEghbA. The Android Application File (APK) can be downloaded at http://zarcrash.x10.mx/OrderEliminationc.apk.

  16. 16.

    The questionnaire can be found at http://zarcrash.x10.mx/OrderEliminationGoogleForm.pdf.

  17. 17.

    The application distributed to participants has been configured with the ray length set to 100 m, the FoV angle set to 28° and a game space of 250,000 m2 (square with side length of 500 m) while the mobile devices used by players varied in technical specifications (Sony Xperia S, Samsung (Galaxy S4 Mini, Beam, Note 4, S4), Nexus (4 & 6), F&U Tablet and Motorola Moto G 2nd Generation).

References

  1. Billinghurst, M., Clark, A., Lee, G.: A survey of augmented reality. Found. Trends Hum. Comput. Inter. 8, 73–272 (2014)

    Article  Google Scholar 

  2. Tian, Y., Long, Y., Xia, D., Yao, H., Zhang, J.: Handling occlusions in augmented reality based on 3D reconstruction method. Neurocomputing 156, 96–104 (2015)

    Article  Google Scholar 

  3. Tian, Y., Guan, T., Wang, C.: Real-time occlusion handling in augmented reality based on an object tracking approach. Sensors 10, 2885–2900 (2010)

    Article  Google Scholar 

  4. Hayashi, K., Kato, H., Nishida, S.: Occlusion detection of real objects using contour based stereo matching. In: International Conference on Augmented Tele-Existence, pp. 180–186. ACM (2005)

    Google Scholar 

  5. Kim, H., Yang, S.-J., Sohn, K.: 3D reconstruction of stereo images for interaction between real and virtual worlds. In: IEEE and ACM International Symposium on Mixed and Augmented Reality, pp. 169–176. IEEE (2003)

    Google Scholar 

  6. Ohta, Y., Sugaya, Y., Igarashi, H., Ohtsuki, T., Taguchi, K.: Client/server depth sensing for see-through head-mounted displays. Presence: Teleoperators Virtual Environ. 11, 176–188 (2002)

    Article  Google Scholar 

  7. Dey, A., Sandor, C.: Lessons learned: Evaluating visualizations for occluded objects in handheld augmented reality. Int. J. Hum Comput Stud. 72, 704–716 (2014)

    Article  Google Scholar 

  8. Clark, A., Piumsomboon, T.: A realistic augmented reality racing game using a depth-sensing camera. In: 10th International Conference on Virtual Reality Continuum and Its Applications in Industry, pp. 499–502. ACM (2011)

    Google Scholar 

  9. Benford, S., Crabtree, A., Flintham, M., Drozd, A., Anastasi, R., Paxton, M.: Can You See Me Now? ACM Trans. Comput. Hum. Interact. 13, 100–133 (2006)

    Article  Google Scholar 

  10. Hodson, H.: Google’s ingress game is a gold mine for augmented reality. New Sci. 216, 19 (2012)

    Google Scholar 

  11. Herbst, I., Braun, A.-K., McCall, R., Broll, W.: TimeWarp: interactive time travel with a mobile mixed reality game. In: 10th International Conference on Human Computer Interaction with Mobile Devices and Services, pp. 235–244. ACM (2008)

    Google Scholar 

  12. Fischer, J., Lindt, I., Stenros, J.: Final Crossmedia Report (part II) – Epidemic Menace II Evaluation report (2006)

    Google Scholar 

  13. Kasapakis, V., Gavalas, D.: Blending history and fiction in a pervasive game prototype. In: 13th International Conference on Mobile and Ubiquitous Multimedia, pp. 116–122. ACM (2014)

    Google Scholar 

  14. Schroeder, J.: AndEngine for Android game development cookbook. Packt Publishing Ltd., Birmingham (2013)

    Google Scholar 

  15. Cheok, A.D., Sreekumar, A., Lei, C., Thang, L.M.: Capture the flag: mixed-reality social gaming with smart phones. IEEE Pervasive Comput. 5, 62–63 (2006)

    Article  Google Scholar 

  16. Wetzel, W., Blum, L., McCall, R., Oppermann, L., Broeke, T.S., Szalavári, Z.: Final prototype of TimeWarp application (2009)

    Google Scholar 

  17. Kasapakis, V., Gavalas, D.: Pervasive gaming: status, trends and design principles. J. Netw. Comput. Appl. 55, 213–236 (2015)

    Article  Google Scholar 

  18. Behzadan, A.H., Kamat, V.R.: Scalable algorithm for resolving incorrect occlusion in dynamic augmented reality engineering environments. Comput. Aided Civ. Infrastruct. Eng. 25, 3–19 (2010)

    Article  Google Scholar 

  19. Fischer, J., Huhle, B., Schilling, A.: Using time-of-flight range data for occlusion handling in augmented reality. In: IPT/EGVE, pp. 109–116 (2007)

    Google Scholar 

  20. Zhu, J., Pan, Z., Sun, C., Chen, W.: Handling occlusions in video-based augmented reality using depth information. Comput. Animation Virtual Worlds 21, 509–521 (2010)

    Article  Google Scholar 

  21. Yang, T., Pan, Q., Li, J., Li, S.Z.: Real-time multiple objects tracking with occlusion handling in dynamic scenes. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 970–975. IEEE (2005)

    Google Scholar 

  22. Lepetit, V., Berger, M.-O.: Handling occlusion in augmented reality systems: a semi-automatic method. In: IEEE and ACM International Symposium on Augmented Reality, pp. 137–146. IEEE (2000)

    Google Scholar 

Download references

Acknowledgement

This work has been supported by the CIP-ICT-PSP-2013-2017 Programme under grant agreement no. 621133 (HoPE - “Holistic Personal public Eco-mobility”).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Damianos Gavalas .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this paper

Cite this paper

Kasapakis, V., Gavalas, D., Galatis, P. (2016). An Efficient Geometric Approach for Occlusion Handling in Outdoors Augmented Reality Applications. In: De Paolis, L., Mongelli, A. (eds) Augmented Reality, Virtual Reality, and Computer Graphics. AVR 2016. Lecture Notes in Computer Science(), vol 9768. Springer, Cham. https://doi.org/10.1007/978-3-319-40621-3_30

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-40621-3_30

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-40620-6

  • Online ISBN: 978-3-319-40621-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics