Abstract
When a natural disaster occurs, there is often significant damage to vitally important infrastructure. Repair crews must quickly locate the structures with the most damage that are in need of immediate attention. These crews need to determine how to allocate their resources most efficiently to save time and money without having to assess each area individually. To streamline this process, drone technology can be used to take photographs of the affected areas. From these photographs, three dimensional models of the area can be constructed. These models can include point clouds, panoramas, and other three-dimensional representations. This process is called photogrammetry. The first step in constructing a three-dimensional model from two dimensional photographs is to detect key features that match throughout all the photos. This is done using David Lowe’s Scale Invariant Feature Transform (SIFT) algorithm which detects the key features. Pairwise matches are then computed by using a k nearest neighbor algorithm to compare all the images one pair at a time finding pixel coordinates of matching features. These pixel matches are then passed to an algorithm which calculates the relative camera positions of the photos in a 3D space. These positions are then used to orient the photos allowing us to generate a 3D model. The purpose of this research is to determine the best method to generate a 3D model of a damaged area with maximum clarity in a relatively short period of time at the lowest possible cost; therefore, allowing repair crews to allocate resources more efficiently.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 60(2), 91–110 (2004)
Introduction to SIFT (Scale-Invariant Feature Transform). (n.d.). Retrieved from OpenCV website. https://docs.opencv.org/3.3.0/da/df5/tutorial_py_sift_intro.html
Wu, C. http://ccwu.me
What is ASPRS? Retrieved from ASPRS website. https://www.asprs.org/organization/what-is-asprs.html
Point Cloud Data: Retrieved from U.S. Naval Academy website. https://www.usna.edu/Users/oceano/pguth/md_help/html/pt_clouds.htm
Shervais, K.: Structure from Motion Introductory Guide. Retrieved from UNAVICO website. https://www.unavco.org/education/resources/modules-and-activities/field-geodesy/module-materials/sfm-intro-guide.pdf
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Eckert, G., Cassidy, S., Tian, N., Shabana, M.E. (2020). Using Aerial Drone Photography to Construct 3D Models of Real World Objects in an Effort to Decrease Response Time and Repair Costs Following Natural Disasters. In: Arai, K., Kapoor, S. (eds) Advances in Computer Vision. CVC 2019. Advances in Intelligent Systems and Computing, vol 943. Springer, Cham. https://doi.org/10.1007/978-3-030-17795-9_22
Download citation
DOI: https://doi.org/10.1007/978-3-030-17795-9_22
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-17794-2
Online ISBN: 978-3-030-17795-9
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)