Stereo-vision-based crop height estimation for agricultural robots

https://doi.org/10.1016/j.compag.2020.105937Get rights and content

Highlights

  • A stereo camera system captures accurate crop height measurements.

  • Disparity and depth maps determine the edges of regions of interest (ROI).

  • Crop regions are segmented using the edges without pre-labelling.

  • The system detects the target crop region even when objects overlapped in images.

  • Estimated crop heights showed strong linear correlations with actual crop heights.

Abstract

The objective of this study was to develop a machine-vision-based height measurement system for an autonomous cultivation robot. The system was developed with a simple stereo camera configuration to facilitate practical field applications and was used to acquire accurate height measurements of various field crops. The acquired stereo images were converted to disparity maps through stereo matching, and the disparity of each pixel was calculated to determine the depth of the distance between the camera and the crop. Depth maps were used to determine the edges of regions of interest (ROI) and the crop regions were segmented using the edges located in the expected crop region closest to the camera without using any additional labelling. The crop height was calculated using the highest points in the ROI. This approach was tested on five crops, and the results showed that the system could detect the target crop region even when objects overlapped in the acquired images. Furthermore, the crop heights estimated with the developed system showed strong agreement with actual crop heights measured manually, with the R2 ranging from 0.78 to 0.84. These results indicate that the developed algorithm is capable of measuring crop heights in various ranges for agricultural robot applications.

Introduction

The global agricultural robot market was estimated to be $4.1 billion in 2017 and is expected to increase steadily to approximately $25.2 billion in 2025 (Market research report, 2019). Market growth in this sector has accelerated because of recent agricultural issues, such as aging-related reductions in the labor force and a projected shortfall in the amount of food available for the global population in 2050. These issues have pushed the agricultural industry to develop advanced autonomous farming products to reduce the labor force, enhance productivity, and manage crops more efficiently. Specifically, automatic monitoring or scouting robots that can survey crop growth without human resources have been studied extensively because crop growth monitoring is among the most frequent and time-consuming processes in farm operations (Chou et al., 2019).

Crop height is an especially important phenotypic factor that indicates overall crop growth and is used to predict yield and manage crop cultivation. Crop height is usually observed manually, which is labor-intensive and costly (Hu et al., 2018); hence, several approaches have tried to measure crop height automatically with various sensors. For instance, Zhang and Grift (2012) developed a measurement system for the stem height of Miscanthus giganteus based on Light Detection and Ranging (LiDAR). The accuracy of the developed system was over 92% in terms of crop density and was reported to achieve real-time crop height evaluation. Although the system can collect 3D structural data accurately, the LiDAR used in this work is costly and has portability limitations for in-field farming machines. Ultrasonic sensor-based approaches that is more economical than LiDAR have also been conducted. Chang et al. (2017) developed a robot-based system using an ultrasonic-based plant height measurement system to control the head position of a harvester during blueberry harvesting. The first prototype of the system could estimate blueberry height within an error of 5.7 cm, and the second prototype reduced the error to 1.7 cm. They showed remarkable results; however, the accuracy of ultrasonic sensors is affected by the material’s softness, and such measurements are challenging when objects are covered in an extremely soft fabric that absorbs sound waves, as is often the case with crops. Ultrasonic waves are also unreliable in agriculture because of their sensitivity to temperature and wind and their associated impacts on the measurements over time.

In the last decade, vision-based technologies have received increasing attention because they can be tag-free, offer inexpensive configurations (Carrasco, 2011), and can provide 3D structural information. 3D image-based studies targeting object recognition in agriculture have been conducted with stereo vision and depth cameras (e.g., Microsoft Kinect). For example, 3D structure estimations have been applied to plant phenotyping (Andújar et al., 2016, Jiang et al., 2016, Malekabadi et al., 2019) and crop row mapping (Jay et al., 2015). Among studies specifically targeting crop height, Jiang et al. (2016) developed a tractor-attached system to measure the height of cotton plants using a Kinect-v2, and the results showed that the system could measure the height under real-world field conditions. Andújar et al. (2016) constructed a 3D model of cauliflower plants using a Kinect-v1, and the developed algorithm showed only a 2-cm deviation compared to the actual height. Malekabadi et al. (2019) obtained stereo images and calculated disparity maps of trees by foliage density and canopy shape, and the developed algorithm could calculate height with errors of less than 7% for both elliptical and conical trees. However, although these studies have advanced crop height estimation technology, they were conducted for only one type of crop and in a specific environment or laboratory condition with controlled light intensity. In particular, Kinect uses structured-light (v1) or time-of-flight (v2) based depth estimation, and the use of infrared radiation range wavelengths makes it difficult to implement under natural light conditions.

Our purpose is to estimate crop heights under uncontrolled field conditions, and the target task is focused on cultivation, not phenotyping. The aforementioned studies were generally performed for phenotypic analyses with individual small sized crops, and these approaches have limitations when applied to automatic cultivation. Images were acquired using cameras oriented towards the nadir with a stable background, and the region of interest (ROI) in each work was predetermined using GPS or markers. The former is not practical for a forward view of automatic cultivation; the latter is cumbersome preliminary work, making end-to-end automation difficult. For use in practical cultivation, such a system must be able to acquire crop images in the forward travel direction and to distinguish crop regions from backgrounds such as soil, weeds, and other crops with similar phenotypes, without additional labelling. Therefore, with the final goal of designing an agricultural robot for cultivation, this study focused on stereo vision-based crop height estimation of various upland crop types under uncontrolled field conditions, with images acquired in the forward direction and gathered from overlapping crops at various distances. Our system can provide the representative height of a target crop region automatically in an end-to-end process during crop cultivation, and it can also contribute to realizing precision agriculture by real-time crop height monitoring with spatial matching.

Section snippets

Stereo image matching

Stereo cameras used to acquire crop images have binocular stereo vision, as illustrated in Fig. 1, implemented by two image sensors with parallel directions. Thus, the distance between the camera and the object can be calculated by Eq. (1), where the distance of the target point (p) projected on the left (xL) and right (xR) image sensors is calculated with depth information acquired from the disparities between the two images in the horizontal (x-axis) pixel distance of the target point (Adi

Results of stereo matching

The representative disparity maps for each crop are shown in Fig. 6. The disparity maps are shown in gray scale, and the intensity of each pixel depends on the distance, such that the closer to white the color is, the closer the object is to the camera. The black band of the disparity map is a region where the left and the right image do not match due to the horizontal shift between images.

The Chinese cabbage images showed high-density crop regions in which the leaf area was larger and leaves

Conclusion

In this study, stereo-vision-based automatic crop height estimation was demonstrated with a system that can be practically applied in an agricultural robot or a conventional farm machine. The developed system has a simple configuration that includes a stereo camera and a single board computer. This system does not require preliminary labelling, and it was shown to successfully measure the height of field crops in front of the machinery on which it was mounted during cultivation. The proposed

CRediT authorship contribution statement

Wan-Soo Kim: Investigation, Validation, Writing - original draft, Writing - review & editing. Dae-Hyun Lee: Conceptualization, Software, Supervision, Writing - original draft, Writing - review & editing. Yong-Joo Kim: Funding acquisition, Project administration. Taehyeong Kim: Conceptualization, Software. Won-Suk Lee: Conceptualization, Writing - review & editing. Chang-Hyun Choi: Conceptualization, Supervision.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgements

This research was supported by the Agricultural, Forestry, Food Research Center Support Program (Project No.: 714002-07), Ministry of Agricultural, Food and Rural Affairs. It was also supported by the Korea Institute of Planning and Evaluation for Technology in Food, Agriculture, Forestry (IPET) through the Advanced Production Technology Development Program, funded by the Ministry of Agriculture, Food and Rural Affairs (MAFRA) (319041-03).

Cited by (68)

View all citing articles on Scopus
View full text