Skip to main content
Log in

Simulation in real conditions of navigation and obstacle avoidance with PX4/Gazebo platform

  • Original Article
  • Published:
Personal and Ubiquitous Computing Aims and scope Submit manuscript

Abstract

In the future, UAVs should be a part of the IoT ecosystems. Integration of sensors onboard allows to enrich information stored in the cloud and, at the same time, to improve the capacities of UAVs. Developing new sensors and the integration in UAV architecture could improve control functions. Design of future UAV systems requires from advanced tools to analyze the system components and their interaction in real operational conditions. In this work, authors present an approach to integrate and evaluate a LIDAR sensor and the capacity for improving navigation and obstacle avoidance functions in simulated situations using a real UAV platform. It uses available software for mission definition and execution in UAVs based on PixHawk flight controller and peripherals. The proposed solution (a general method that could be used to integrate other kind of sensors) shows physical integration of the main types of sensors in UAV domain both for navigation and collision avoidance, and at the same time the use of powerful simulation models developed with Gazebo. Some illustrative results show the performance of this navigation and obstacle avoidance function using the simulated sensors and the control of the real UAV in realistic conditions.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1.
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18
Fig. 19
Fig. 20
Fig. 21
Fig. 22
Fig. 23

Similar content being viewed by others

Notes

  1. NuttX Real-Time Operating System, http://nuttx.org/

  2. http://docs.px4.io/en/sensor/LIDAR_lite.html

References

  1. Garcia J. Molina J.M., J. Trincado. Analysis of real data with sensors and estimation outputs in configurable UAV platforms. Sensor Data Fusion: trends, solutions and applications 2017. IEEE Bonn, Germany, 10–12 Oct. 2017

  2. Groves PD (2015) Navigation using inertial sensors. IEEE AES Magazine 30(2):42–69

    Article  Google Scholar 

  3. Britting KR (2010) Inertial navigation systems analysis. Artech House

  4. Farrel JA (2008) Aided navigation: GPS with high rate sensors. McGraw-Hill, New York

    Google Scholar 

  5. Choi H, Kim Y (2014) UAV guidance using a monocular-vision sensor for aerial target tracking. Control Eng Pract 22:10–19

    Article  Google Scholar 

  6. Torres-González A, Martínez-de Dios JR, Ollero A (2017) Robot-beacon distributed range-only SLAM for resource-constrained operation. Sensors 17(4):903

    Article  Google Scholar 

  7. Ferrick A., Fish J., Venator E. and Lee G.S. UAV obstacle avoidance using image processing techniques Technologies for Practical Robot Applications (TePRA), 2012 IEEE international conference on 23–24 2012

  8. Fasano G, Accado D, Moccia A, Moroney D (2016) Sense and avoid for unmanned aircraft systems. IEEE Aerosp Electron Syst Mag 31(11):82–110

    Article  Google Scholar 

  9. Rankin G., Tirkel A., Leukhin A. Millimeter wave array for UAV imaging. MIMO Radar Radar Symposium (IRS), 2015 16th International 24–26 June 2015

  10. Gualda D, Ureña J, García JC, García E, Alcalá J (2019) Simultaneous calibration and navigation (SCAN) of multiple ultrasonic local positioning systems. Information Fusion 45:53–65

    Article  Google Scholar 

  11. Fasano G, Accardo D, Tirri AE, Moccia A (2016) Experimental analysis of onboard non-cooperative sense and avoid solutions based on radar, optical sensors, and data fusion. IEEE Aerosp Electron Syst Mag 31(7):6–14

    Article  Google Scholar 

  12. Liu Z, Foina AG (2016) An autonomous quadrotor avoiding a helicopter in low-altitude flights. IEEE Aerosp Electron Syst Mag 31(7):30–39

    Article  Google Scholar 

  13. J. García, J.M. Molina. Analysis of sensor fusion solutions for UAVs, Conference of the Spanish Association for Artificial Intelligence (CAEPIA). 2018. Granada, Spain. 23-26 2018

  14. Meier L., PX4 Development Guide (online). Available: https://dev.px4.io/en/

  15. Trawny, Nikolas, and Stergios I. Roumeliotis. Indirect Kalman filter for 3D attitude estimation. University of Minnesota, Dept. of Comp. Sci. & Eng., Tech. Rep 2 (2005)

  16. Sola, Joan. Quaternion kinematics for the error-state Kalman filter. arXiv preprint arXiv:1711.02508 (2017)

  17. G. Cai., B. M. Chen, Unmanned rotorcraft systems, Chapter 2 - Coordinate System and Transformation, 2011

  18. https://es.mathworks.com/help/robotics/examples/perform-co-simulation-between-simulink-and-gazebo.html(accessed Oct 19)

  19. https://es.mathworks.com/help/supportpkg/px4/ug/enable-mavlink-px4.html(accessed Oct 19)

  20. T. R. Tuinstra, Range and velocity disambiguation in medium PRF radar with the DBSCAN clustering algorithm, , de : aerospace and electronics conference (NAECON) and Ohio innovation summit (OIS) Dayton, OH, USA, 2016

  21. Y. Zhao Y. Su, Vehicles detection in complex urban scenes using Gaussian mixture model with FMCW radar, IEEE Sensors J , vol. 17, n° 18, pp. 5948–5953, 2017

  22. MAVLINK common message set, https://mavlink.io/en/messages/common.html

  23. Fasano G, Accado D, Moccia A (2016) Sense and avoid for unmanned aircraft systems. IEEE Aerospace and Electronic Systems Magazine:82–110

  24. Kotegawa T (2016) Proof-of-concept airborne sense and avoid system with ACAS-XU flight test. IEEE Aerosp Electron Syst Mag 31(7):53–62

    Article  Google Scholar 

  25. Schulman J et al (2013) Finding locally optimal, collision-free trajectories with sequential convex optimization. Robotics: science and systems 9(1)

  26. Martí E, García J, Escalera A, Molina JM, Armingol JM (2012) Context-aided sensor fusion for enhanced urban navigation. Sensors 2012:16802–16837

    Article  Google Scholar 

  27. Layh T, Demoz GE (September 2017) design for graceful degradation and recovery from GNSS interruptions. IEEE AESS Magazine Vol 32(N9):4–17

    Article  Google Scholar 

Download references

Acknowledgments

Authors are grateful to Enrique Martínez Martínez and Ivan Serrano Ruiz, Graduate Engineers at UC3M in 2018 for their helpful collaboration in this work.

Funding

This work was funded by public research projects of Spanish Ministry of Economy and Competitiveness (MINECO), references TEC2017-88048-C2-2-R, RTC-2016-5595-2, RTC-2016-5191-8 and RTC-2016-5059-8.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jesús García.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

LIDAR Simulation in Gazebo Environment

LIDAR Simulation in Gazebo Environment

This section illustrates the use of Gazebo environment to generate the sensor input for navigation and obstacle avoidance function. As mentioned, the Iris quadrotor model developed by 3D Robotics is available for direct use from QGround platform and PX4 flight control, containing models of inertial sensors and GPS receiver as output data, and receiving input signals from flight controller to drive the four engines and simulate its dynamics. As an example of configuration for the simulation in Gazebo, the available model includes simulation of GPS noise, with a behavior similar to that typically found in real systems, a feature useful for testing applications that might be impacted by noise in positioning.

So, the GPS noise is enabled if the target vehicle’s SDF file contains a value for the gpsNoise element (i.e., it has the line: <gpsNoise> true </ gpsNoise>). It is enabled by default in many vehicle SDF files, including the quadrotors solo.sdf and iris.sdf. To enable/disable GPS noise the following steps should be carried out:

figure a

Analogously, the model of the LIDAR sensor has been developed with “.sdf” file which reflects the logic behind the sensor’s rays and the information produced. After the definition, the available Iris drone model was extended with a LIDAR sensor coupled in its upper part. The basic specification of Gazebo file is as follows:

figure b

To understand in a more detailed way the operation and the configuration for this sensor, the corresponding .sdf file for LIDAR sensor containing the information referring to the sensor is shown below, including the physical parameters of sensor (location, mass, geometry, etc.), configuration of sensor range and angular coverage, and a facility to visualize the data:

figure c

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

García, J., Molina, J.M. Simulation in real conditions of navigation and obstacle avoidance with PX4/Gazebo platform. Pers Ubiquit Comput 26, 1171–1191 (2022). https://doi.org/10.1007/s00779-019-01356-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00779-019-01356-4

Keywords

Navigation