Next Article in Journal
Robust Heartbeat Classification for Wearable Single-Lead ECG via Extreme Gradient Boosting
Next Article in Special Issue
Research on Pepper External Quality Detection Based on Transfer Learning Integrated with Convolutional Neural Network
Previous Article in Journal
Intuitive Cognition-Based Method for Generating Speech Using Hand Gestures
Previous Article in Special Issue
Opportunities and Possibilities of Developing an Advanced Precision Spraying System for Tree Fruits
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Obstacle Detection System for Agricultural Mobile Robot Application Using RGB-D Cameras

1
Unitem, ul. Kominiarska 42C, 51-180 Wrocław, Poland
2
Faculty of Electronics, Wrocław University of Science and Technology, Wybrzeże Wyspiańskiego 27, 50-370 Wrocław, Poland
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5292; https://doi.org/10.3390/s21165292
Submission received: 30 June 2021 / Revised: 30 July 2021 / Accepted: 30 July 2021 / Published: 5 August 2021
(This article belongs to the Special Issue Sensing Technologies for Agricultural Automation and Robotics)

Abstract

:
Mobile robots designed for agricultural tasks need to deal with challenging outdoor unstructured environments that usually have dynamic and static obstacles. This assumption significantly limits the number of mapping, path planning, and navigation algorithms to be used in this application. As a representative case, the autonomous lawn mowing robot considered in this work is required to determine the working area and to detect obstacles simultaneously, which is a key feature for its working efficiency and safety. In this context, RGB-D cameras are the optimal solution, providing a scene image including depth data with a compromise between precision and sensor cost. For this reason, the obstacle detection effectiveness and precision depend significantly on the sensors used, and the information processing approach has an impact on the avoidance performance. The study presented in this work aims to determine the obstacle mapping accuracy considering both hardware- and information processing-related uncertainties. The proposed evaluation is based on artificial and real data to compute the accuracy-related performance metrics. The results show that the proposed image and depth data processing pipeline introduces an additional distortion of 38 cm.

1. Introduction

In complex robotic tasks, such as in agricultural applications, sensing systems are a key element influencing the functionality and effectiveness of job execution. Proper selection of the sensing system is usually made in the light of defined functionalities and requirements established at the robot design stage. In this context, a sensory system based on computer vision has a special place due to its versatility. It can be used in numerous applications, ranging from fruit detection (including size, color, ripeness, etc.) to object tracking and establishing the robot path after scene analysis [1]. For this reason, vision systems are the basis for autonomous robot navigation, since in a large number of applications, paths are marked/defined/detected visually [2]. As an example, the orchard’s robot navigation can be introduced, where the mobile robot needs to detect the path autonomously (e.g., defined as the position in the middle of the row) to accomplish their task (recollection, spraying, crops stand counting, plant phenotyping, etc.) [3,4,5,6]. In agricultural applications, the path detection and navigation strongly depend on the robotic task goals. There are different ways the path is specified/defined, and various constraints affect robot operations, including image recognition and maneuvering—the most typical examples include road detection [7], crop row guidance for agriculture robots [3,8], area detection for lawn movers [9], and obstacle detection and procedures for their avoidance [10].
Simultaneously with the task’s objectives, most autonomous robots need to detect and avoid obstacles to ensure safe operation. Obstacle detection systems can include a large number of sensors and techniques ranging from ultrasonic sensors and lidars to more complex solutions relying on computer vision systems. Due to their flexibility, cameras are universal and widely applicable in tasks where a complex perception system is required [1,5,6]. In this context, physical camera characteristics, such as resolution, field of view, etc., are only part of the overall computer vision system properties. In these systems, most features are defined by the implemented software and data processing algorithm. For this reason, in robotic applications, where the computer vision system is used for navigation purposes (e.g., obstacle detection and mapping), the accuracy error introduced by the processing software should also be considered to improve the localization and navigation effectiveness. The robotic agricultural application considered in this work, lawn mowing, assumes the robot knows only the working area boundary, expressed by the GPS coordinates, and the robot’s main objective is to cover the working area [9,11]. To complete this goal, the robot global path is generated to optimize robot movements and to perform tasks more efficiently. However, in this approach, the robot has no information regarding the obstacles’ positions and needs to detect and avoid them in an autonomous manner using its sensors. Detected obstacles need to be mapped in order to be avoided by the robot. To accomplish this task, the robot’s global path must be divided into smaller parts (e.g., using the idea of moving horizon), known as a local plan. Afterwards, the local planner takes into account the global goals and obstacles and generates the local plan, allowing obstacles to be avoided and simultaneously performing as close as possible to the global path. The described approach is widely used in many agricultural tasks [12,13]. The performance of the job execution depends mainly on the type (static or dynamic) and the number of obstacles, influencing the obstacle mapping precision and, as a consequence, obstacle avoidance performance and effective area coverage. Additionally, for safety reasons, the obstacle’s shape is augmented (“inflated”) with a safety distance that is added to the original dimension to make the avoidance procedure safe [10,14]. To make this navigation strategy efficient (maximize area coverage), obstacles in working areas should be localized with high precision. In practical applications, obstacle localization is usually not very precise, the safety distance is overestimated (to promote safety over performance), and the efficiency in area coverage drops. However, if one can improve the estimation of the required safety distance, the criteria for both safety and performance can be satisfied, while the area coverage can be maximized. This paper describes a novel methodology to determine the obstacle detection accuracy for a RGB-D camera-based vision system. The main objective is to compute the real precision of the system considering both hardware-related and information processing pipeline (software)-related accuracy. The methodology used here exploits virtual and real data to calculate the accuracy error related to the sensor and to the data processing pipeline. In this approach, artificial data are used to determine the average mapping distortion introduced by the software processing. Using this information and computed indexes from the real scene analysis, the accuracy related to the sensor precision can be established. As a consequence, the real accuracy of obstacle mapping is determined, and then the real precision can be considered in the planning algorithm for obstacle avoidance. Thanks to this, the safety margin can be set within real needs, resulting in better planning and an improved ratio between working and covered areas.
The main contribution of the paper is related to a new evaluation method for obstacle detection, i.e., marking obstacles on the map, which is used by the navigation module. To determine the mapping accuracy, a new metric in terms of obstacle detection was defined. The described research has also shown that segmentation metrics are not sufficient for vision/depth systems in terms of obstacle detection and the navigation module, and additional metrics are required to unequivocally determine the mapping performance. The proposed accuracy verification method was tested on sequences collected from real sensors. The experimental evaluation performed confirmed the effectiveness of the introduced method.
This paper is organized as follows: Section 2 presents a detailed analysis of related works. Section 3 summarizes the system architecture used for obstacle detection. In the first place, vision based techniques for obstacle detection are described. Subsequently, it is shown how these obstacles are mapped and handled in a navigation task. The motivation and methodology for accuracy determination in the presented system are also proposed. Section 4 deals with the experimental evaluation of the proposed methodology using artificial and real data sets. Additionally, the verification measures and the procedure for the avoidance system setup, including determined accuracy, are proposed. Finally, conclusions are given in Section 5.

2. Related Work

Movement, navigation, and obstacle avoidance are crucial aspects of all robotic systems, irrespective of robot design, application field, and performed tasks (e.g., [15,16,17].) While most agricultural robots move on fields and navigate their path along crop or orchard rows [3,8,15,18], lawn mowers traverse soft surfaces, require high maneuverability, and navigate through an area which is quite uniform with no natural landmarks to follow. This makes it more challenging to navigate the robot, detect deviations from the desired path, and complicate obstacle detection. As a result, the safe operation of lawn mowers requires various sensing technologies including RGB and depth cameras [19,20].

2.1. Obstacle Detection

Ball et al. [15] developed a vision-based obstacle detection and navigation for a robot operating in the crop field. The robot navigates the area using a fusion of GPS, inertial sensors, and a camera sensor to follow the crop rows. An additional pair of stereoscopic wide angle cameras are used to observe the area 10 m ahead of the robot and detect obstacles. Obstacles are detected by looking for so-called novelty in the cropping image [21]. The method aims to detect the difference between the current and recent image-detecting candidates for obstacles. This is possible for robots operating in fields where crops/stubble are straight and parallel because the majority of crops and stubble can be removed from consideration, simplifying image processing. Unfortunately, this approach generates a number of false positive detections, especially when the robot moves across the crop rows and turns at the headland turning regions. Additionally, it is sometimes difficult to distinguish between obstacles and the ground when the obstacle is close to the ground. Delayed detection and a short range make it challenging for the robot to respond to the detected obstacles correctly.
Fleischmann et al. [22] also developed a system based on stereo vision. The whole classification algorithm is based on splitting the generated point cloud into cells and the further analysis of the cells, including the density and distribution of the points and neighborhood of the cell.
The use of an RGB camera and distance measurement was investigated by Brown et al. [23]. They attempted to use a monocular vision camera and a single range measurement to improve depth estimation in agriculture settings. The goal was to mitigate the ambiguity of the scale that exists when only vision images are considered. The results presented show improvements when run on NYUv2 and KITTI datasets but fail to work correctly on agricultural objects which exhibit fractal properties.
Past articles [19,20] present a platform for the evaluation of human detection in autonomous mowers using RGB, stereoscopic, infrared and 360° cameras, LIDAR, and radar. The result of these works is the performance analysis for human detection in different settings and poses. Additionally, they developed a database of sensor recordings and corresponding annotations that can be used for the evaluation of other obstacle detection algorithms. The database includes recordings and locations of humans, buildings, vehicles, and vegetation.
Christiansen et al. [17] proposed a method for obstacle perception using anomaly detection. They benefit from combining deep convolutional neural network (CNN) with anomaly detection to solve the problem of detecting distant and occluded objects, which might not be recognized as an obstacle by typical CNNs. The presented tests show that the proposed technique enables a human to be detected at longer ranges, up to 90 m, preserving real-time operation.
In the literature, there are more approaches that use deep learning. In [24], the authors fine-tuned the popular AlexNet CNN [25]. The method was examined in terms of the ISO/DIS 18497 norm, which defines a barrel-shaped obstacle. To comply with the standard, the obstacle should be fully detectable by the system. The authors conclude that cameras may also be used as a part of safety systems, showing that precision is more than 90%.
Cameras are not an obligatory part of the obstacle detection system. Other studies show that only range sensors may be used. In [26], researchers utilize LIDAR data. The SVM algorithm based on the point cloud originating from sensors classifies points as obstacles, vegetation, or objects. The accuracy of the classification was 91.6%. Dvorak et al. [27] investigated the ultrasonic sensor in terms of the capability of detecting numerous obstacles and the possibility of finding in an agricultural field a human model, a Dracaena plant, a wooden fence, and more.
For more results on obstacle detection and techniques used in agricultural applications, the reader is advised to peruse the contents of the available survey articles [28,29,30].

2.2. Autonomous Lawn Mower

The article by Inoue et al. [31] investigates the operation of an autonomous lawn mower that is navigated without a global navigation satellite system (GNSS). Instead, they use cameras and IMU to simultaneously locate the mower, map the area (SLAM), and detect and avoid obstacles. This approach allows the mower to be used in areas where the GNSS signal is weak. The proposed approach was verified in a small 20 by 10 m lawn where the robot traveled a predefined path in the forward and backward directions. Using additional landmarks with known positions, the robot positioning accuracy at the end of the path is equal to 0.38 m and is worse when no landmarks are used. The authors report that an important factor of the error is the drift of the robot during turnaround. Results on obstacle detection and avoidance are not presented.
Visual-only obstacle detection for a lawn mower is presented by Franzius et al. [32]. As presented, the use of a low-cost camera and color-based obstacle detection can navigate the mower on a flat lawn, avoiding objects. In this approach, object detection is based on grass segmentation based on hue and saturation components of the image. The parameters of the segmentation are adapted based on the sequence of images to improve the accuracy of segmentation in variable lighting conditions. The robot operates under several assumptions regarding the operational area: the future driving path is on the same plane as the mower, obstacles lay on the ground, and the distance to the obstacle can be estimated from its pixel coordinates on the image. These assumptions eliminate the need for 3D information and simplify image processing and navigation but simultaneously restrict practical applications. As a result, good avoidance rates (above 90%) are achieved only for obstacles that do not stand out high from the ground. For poles, trees, and other plants, the avoidance rate is much smaller—unacceptable in industrial applications.

3. Obstacle Detection System Architecture

This section presents a novel RGB-D camera-based obstacle detection approach for mobile robots. The proposed method consists of four stages. First, the synchronized and aligned pair of an RGB image and a depth image is collected. Following this, the semantic mask of the environment is obtained based on the RGB image. Next, based on the semantic segmentation result and the depth image, the reconstruction of the point cloud occurs. Finally, the points are projected on the 2D occupancy grid [33], determining the position of obstacles on the map. The obtained grid is then utilized by the navigation algorithm. Figure 1 shows a general scheme of the proposed obstacle detection approach implemented using the robot operating system [34]. Each step of the system is described in more detail in the following subsections.

3.1. Autonomous Mower Platform Overview

The autonomous mower platform is a modified lawn mower for professional use, which is available on the market. Additional components were installed to reach the autonomous features, including sensors, actuators, and a computer. This provides a localization accuracy of 2 cm, which is a result of combining the Piksi Multi RTK module, SICK incremental encoders, and Xsens Mti-7 IMU data. Moreover, it is equipped with obstacle detection sensors which are four Intel RealSense D435i RGB-D cameras and two units of RP-Lidar-S1. All computations are done on an Intel Core i7 8700 PC equipped with an Nvidia GeForce 1050Ti and 16 GB of RAM. The mower can move at up to 0.4 m/s which gives a mowing efficiency at 1800 m 2 /ha. Its width, depth, and height are equal to 1.64 m, 1.43 m, and 0.93 m accordingly.
The platform is equipped with a high-pressure hydraulic circuit that regulates and distributes the oil flow. In this application, the hydraulic system is powered by a gasoline engine that drives the hydraulic pump. The gasoline engine operates at a constant revolution rate to deliver constant power. The compressed oil flow is distributed by the hydraulic regulation circuit. In the analyzed scheme, two hydraulic motors are used to rotate the wheels, one for each side of the robotic platform. The wheel direction can be controlled by the solenoid valves, which are activated by binary signals. Depending on the valve configuration, the wheels can operate as follows: both sides in a forward direction, both sides in a backward direction, or each side in opposing directions. The last configuration is used to perform the skid steer maneuver, which allows the platform to execute a zero-radius turn [35].

3.2. Data Acquisition

The autonomous lawn mowing robot platform used in this study (Figure 2) has four Intel RealSense D435i [36] RGB-D cameras, one pair mounted at the front and the second one at the rear. Based on the mower’s direction of movement, only one pair of cameras is running.
The RealSense camera provides two types of images, RGB images and depth images. Examples of each are presented in the first and the second column, respectively, in Figure 3. The depth frame must be aligned with color camera coordinates. Alignment ensures that every pixel in the RGB image corresponds to exactly one pixel in the depth image. This property is then utilized in the labeled point cloud reconstruction step.

3.3. Semantic Image Segmentation

The goal of semantic segmentation is to classify each pixel in an RGB image into one of the given classes. This allows understanding the image on the pixel level. In the proposed approach, the set of predefined classes can be divided into two categories: one representing the working area of the robot and the other representing obstacles. In the case of autonomous lawn mowers [35,37], the class representing the working area is grass, and the example of the obstacle class is person. It is assumed that every pixel that is recognized as the other class is an obstacle. The list of classes can be adapted to different applications.
To solve the semantic segmentation task, two models were trained: the first one parses the scene into a grass/obstacle mask, and the second one recognizes grass, obstacles, and people. The chosen models’ architecture was the DeepLabv3+ [38] network with the pre-trained ResNet-34 [39] as the feature extractor, whereby the models vary in the last layer: the model grass/obstacle uses the Sigmoid activation function and the model grass/obstacle/person applies SoftMax over features to each spatial location. The segmentation results in a one-channel 448 × 448 mask image where pixel values represent the mower’s environment as follows: 0 implies class other, 1 indicates grass, and for the more complex model, the value 2 indicates the presence of the person. The mask is then up-sampled to match the depth image size. The models were trained using 52,502 pairs of RGB images and the corresponding image segmentation masks. The third column of Figure 3 shows the example results of semantic segmentation using the grass/obstacle/person model.

3.4. Labeled Point Cloud Reconstruction

To obtain the labeled point cloud, the results of semantic segmentation and depth image are fused. This uses the depth to color alignment property described in Section 3.2. The point cloud consists of the XYZL points, which store the information about the coordinates of the point in the 3D world and contain the class label assigned during semantic segmentation. The point coordinates are computed as
x y z = D ( u , v ) 1000 K 1 u v 1 ,
where x , y , z are the real world coordinates, u , v are the image coordinates, D ( u , v ) is the depth value from the depth image in the point u , v , and K is the camera’s intrinsic matrix. The constant 1000 is applied to obtain the coordinates’ result in meters. An example of the point cloud reconstruction is shown in Figure 4. The resulting point cloud is then filtered and downsampled using the Point Cloud Library [40].

3.5. Obstacle Mapping

The final stage of the pipeline is the occupancy grid update. The occupancy grid (map) is used to represent a mobile robot workspace as a discrete grid [33]. The update procedure projects the [ x , y ] coordinates of points from the labeled point cloud onto the map cell’s coordinates ( x map , y map ) :
x map y map = x r + x offset y r + y offset ,
where x offset and x offset are offsets related to map origin and dimensions, and r is the map resolution in meters. The map update has two steps. The process starts with the map clearing of points which were classified as grass. Then, obstacle/person points are added to the map.
In order to include the result of the detection system in the navigation pipeline, the costmap layer, which is the occupancy grid representation in the ROS [34] library, was implemented using the framework proposed by Lu et al. [14].

4. Experiment Overview and Results

As was already mentioned in Section 3, our navigation module is highly dependent on the map, which is represented as an occupancy grid [33]. To finish mowing tasks successfully, reliable map updates are obligatory. The whole vision pipeline processing leads to adding or removing obstacles on the map. Therefore, the vision system evaluation should be considered from a map perspective. In the literature, numerous papers are available regarding obstacle detection systems for mobile robots including [15,17,32,41,42]. However, no article examines the proposed methods in terms of the generation of maps that contain information about working area, unknown places, and obstacle position.
In comparison with other evaluation methods which only test obstacle detection performance on images, i.e., [17,43], it has been noticed that our approach requires manual labeling of the point clouds, which is time consuming and may be affected by errors. However, such an approach allows obstacle detection systems to be compared in a 3D environment where navigation occurs. Moreover, our study has shown that image-based metrics are not sufficient to describe the whole obstacle detection system.
Our experiments could be divided into two separate categories. The goal of the first experiment was to investigate the image segmentation performance and calculate standard metrics for the classification problem, such as accuracy, F1, precision, recall, and specificity [44]. Then, based on point clouds, which were created from depth images, maps were generated. Maps may be considered as grayscale images and compared with reference maps using the same metrics as for the results of semantic segmentation. However, this evaluation method does not provide any information about the geometric properties of the errors, which are crucial to evaluate the safety of the navigation. For this reason, we propose a new metric d min - obst (minimum distance to obstacle), which is calculated for each point on the map that belongs to an FN (false negative) set, i.e., a point that is incorrectly classified as grass while it belongs to an obstacle. For a given FN point p f n , the metric d min - obst is calculated as
d min - obst = min t p i ( p f n p t p i 2 ) ,
where the point p t p i is the map cell correctly classified as the obstacle, T P is the set of all points correctly classified as obstacles, and t p i T P . In the evaluation, we report the statistics of d min - obst values for all FN points on the map.
The idea behind the metric is straightforward. It calculates the Euclidean distance between the false negative map cell and the closest true positive. Figure 5 presents a graphical interpretation of the d min - obst . The proposed metric can be used to estimate a margin for the robot to safely pass an obstacle.
In our research, we consider every pair of RGB and depth images as an input to the obstacle detection system that results in a standalone map. We use these maps to evaluate the quality of the proposed method. The results of the research are discussed in detail in Section 4.2 and Section 4.3.

4.1. Datasets Description

Our tests were conducted on two different datasets, i.e., synthetic and real world sets. The former is a multimodal, artificial dataset of Enclosed garDEN scenes (EDEN) [45]. The latter contains actual images and depth maps recorded by cameras mounted on the autonomous mower.

4.1.1. Synthetic Dataset

We performed our tests using the subset of the EDEN (Enclosed garDEN) dataset, which contained images from 20 different synthetic gardens. For each scene, there are five sequences of images, which were generated in different lighting conditions. The dataset provides various modalities, but in this research, we used only RGB, depth, and semantic segmentation images. We used data obtained from the left camera. In total, 49,762 sets of data were used to evaluate the proposed obstacle detection system. The detailed structure of the test dataset is presented in Table 1.

4.1.2. Real-World Dataset

None of the synthetic datasets can be used to test the proposed solutions. Real-world application requires a realistic dataset which allows carrying out tests in diversified environments and in the presence of noise. Such an evaluation leads to conclusions about the advantages and disadvantages of the examined methods. This is why the second dataset was prepared. It consists of four sequences recorded at different places: boulevard, backyard, and park. An example of each sequence is shown in Figure 3. Each sequence contains two groups of data. The first group is dedicated to the image segmentation task. It has both RGB images and depth maps captured with Intel RealSense D435i camera with VGA resolution. Images were manually labeled using the Computer Vision Annotation Tool (CVAT), which is a free, open-source, browser-based application. It provides convenient annotation instruments [46]. The second group is intended to test obstacle detection on maps. For each sequence, five images were chosen, and based on these, the point clouds were generated. Then, reference point clouds were prepared using the RViz Cloud Annotation Tool [47]. Finally, reference maps were derived from the annotated point clouds. An example of a reference point cloud and reference map is shown in Figure 6.

4.2. Image Segmentation Results

The first stage of the research was to measure the performance of the two classification models on the images. The former model detects two classes, other and grass, while the more advanced one can also extract people. They were examined in terms of the following metrics: accuracy, F1, precision, recall, and specificity.
Table 2 and Table 3 contain results for the synthetic dataset and grass/obstacle model. In almost every illumination condition, the accuracy is higher than 98%. Only one exception can be observed for sunset lighting conditions, where the accuracy is slightly below 98%. The results presented in Table 2 indicate that different illumination conditions do not affect the classification performance significantly, which shows that our model is robust to lighting conditions. Regarding class-dependent metrics, the grass/obstacle model performs better for the other class in terms of F1, precision, and recall. These values are close to 99% for the other class and 91% for grass.
The results for the real-world dataset for both models are presented in the next two tables. Table 4 gathers calculated metrics for the grass/obstacle model, whereas Table 5 summarizes tests for grass/obstacle/person. The accuracy is higher than 96%. The only exception is the park sequence with 87.36%. This is the most challenging scene. The people are on the scene, but they are far away from the camera. Another difficulty is that the grass and footpath boundaries are not clear. This may explain the lower metrics in this case. However, the more advanced model copes with that sequence very well. Tests on real-life images collected on an autonomous mower in various lighting conditions have shown that we are able to correctly detect obstacles with high accuracy of 95.2%. This proves that our models are robust to harsh outdoor environments.

4.3. Map Generation Results

We consider the evaluation of the map generation process in two ways. In the first approach, we assume that the map is a grayscale image in which each pixel represents one of the classes. In the application of an autonomous lawn mower, these classes are grass, obstacle, person, and unknown. Depending on the model chosen for image segmentation, the map uses a proper subset of those classes. The label unknown indicates that the sensor has not covered this area. The quality of the generated map is obtained by comparison with the reference map and is described using the following metrics: accuracy, F1, precision, recall, and specificity. The second approach is more spatially oriented. It is used to analyze the distance from the map cell that is not recognized as an obstacle (although it contains an obstacle) to the closest point correctly recognized as obstacle using the d min - obst metric proposed in (3).
Table 6 and Table 7 show the results acquired for synthetic data. In the first, we can observe a significant performance loss of the classification executed directly on the maps compared to the results obtained for image segmentation (Table 3). This occurs for both resolution values; however, the maps generated in 0.05 m resolution are better at preserving information about objects’ locations. It can be also noticed that the recalls of the grass class, for both 0.05 m and 0.1 m resolutions, are considerably lower than other metrics. This observation reveals that map cells containing grass only are more likely to be misclassified as an obstacle. However, high obstacle class recall, above 97%, shows the robustness of the obstacle misclassification of the proposed obstacle detection system. This property is desirable from the safety system perspective. Despite the differences in classification metrics for the maps of various resolutions, the d min - obst measure shows that for resolutions of 0.05 m and 0.1 m, the system works with similar performance.
The unknown rows in Table 6 represent the map cells that have not been covered by sensors. Because this is an artificial dataset, there is no data noise, and for that reason, it is not possible to mark an unknown cell as an obstacle or grass.
In the case of the real world dataset, we need to deal with noise and distortion both in RGB and depth images. The results for the first approach performed on the data processed by the grass/obstacle/person model are presented in Table 8 and Table 9 for 10 cm and 5 cm resolution, respectively. For Boulevard 1 and Park sequences, the map generation with the lower resolution (10 cm) has significantly better performance than the other one in terms of obstacle detection. This shows that the resolution parameter can function as a filter. However, in comparison with Table 5, significantly worse results were obtained. In particular, for the Park sequence, the precision of obstacle detection decreased from 92.5% to only 34.11% for 10 cm resolution and 26.13% for 5 cm. This proves that the obstacle detection system cannot be examined using RGB images only. It is also worth noting that the depth image noise is revealed by the metrics’ values for class unknown.
Table 10 contains the evaluation of the real world map results using the d min - obst metric. The analysis shows again that for the sequence Park, our system obtains the worst results. The average d min - obst for this sequence is 2.74 m and 3.13 m for a map resolution of 0.1 m and 0.05 m, respectively. However, this is the most challenging sequence from our dataset, and it appears that the model cannot detect the pathway in the scene as an obstacle correctly. The use of the d min - obst revealed the difficulty in detecting the obstacles in the Park scene which was not evident regarding the image-based evaluation. Figure 7 presents an example of a poorly detected dirt trail. We manage to get very good results for the sequence Backyard, but it is the most static scene from the real world dataset. Sequences Boulevard 1 and Boulevard 2 present dynamic scenes depicting humans. The first shows a woman walking right in front of the robot. The second one presents a biker stopping in front of the mower and riding away. The scene is more challenging because not only is the person moving, but the bike is also present, which represents the class other. For the second sequence, we achieved a mean value of the d min - obst equal to 0.38 m for the map resolution 0.1 m. This indicates the minimum value that should be added to the safety margin to navigate in a dynamic environment.

5. Conclusions

In this work, a methodology for accuracy determination in obstacle detection and mapping using RGB-D camera-based vision was proposed. In the proposed approach, artificial and real multimodal data containing a photo and 3D representation of the scene are used. These data are explored in the vision processing pipeline, where the first RGB picture is analyzed to detect obstacles in the working area. At this stage, it was shown that the scene segmentation obtains an average accuracy of 98.11%. Later on, the obstacles are matched with depth information for mapping the obstacle on the 3D map. Two data sources allowed the analysis to be performed in two parts, related to data processing issues (interpolation, resampling, point cloud size reduction, filtering, etc.) and hardware (sensor). It was determined that the average accuracy error introduced by the information processing is around 38 cm. The obtained value is considered in obstacle avoidance planning to establish the safety margin that is added to the detected obstacle radius. In this way, the safety margin can be established within real needs and not overestimated (the usual situation in practical implementations). Therefore, using the proposed approach in agricultural applications can lead to significant improvements in operational performance.
The proposed methodology can be used in other autonomous systems where the obstacle detecting system is based on RGB-D cameras. In this way, the safety margin can be adjusted to real requirements; as a consequence, more efficient path following and obstacle avoidance can be achieved. Finally, the task execution time can be reduced by limiting the robotic platform maneuvering, simultaneously keeping the required safety distance. The research also shows that obstacle detection is a complex system, and its performance cannot be assessed only based on image segmentation tests. The study proves that the final results of the obstacle detection system are affected by image segmentation, depth measurement errors, and point cloud reconstruction. Consequently, a comprehensive evaluation of the obstacle detection system needs to include navigation and mapping systems.
Future work will concentrate on increasing the system robustness in harsh, unknown environments, while maintaining long-term safety by using the d min - obst metric proposed in this paper. We also hope to look into the possibility of performing the scene segmentation step using both RGB and depth image data. Additionally, we are in the process of enhancing obstacle recognition performance for the group of obstacles that were identified in this research as challenging to recognize for the models used, such as dirt trails.

Author Contributions

Conceptualization, M.S., M.O., M.N. and A.P.; methodology, M.S. and M.O.; software, M.S., M.O., K.S., D.K. and M.P.; validation, M.S., M.O. and K.S.; data curation, M.S. and M.O.; writing—original draft preparation, M.S., M.O., M.N. and A.P.; writing—review and editing, M.N. and A.P.; visualization, M.S. and M.O.; supervision, A.P. and M.N.; funding acquisition, M.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research and the APC was funded by The National Centre for Research and Development, grant number POIR.01.01.01-00-1069/18.

Data Availability Statement

EDEN dataset: https://lhoangan.github.io/eden/ (accessed on 3 August 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hongkun, T.; Tianhai, W.; Yadong, L.; Xi, Q.; Yanzhou, L. Computer vision technology in agricultural automation—A review. Inf. Process. Agric. 2020, 7, 1–19. [Google Scholar] [CrossRef]
  2. VanQuang, N.; Heungsuk, K.; SeoChang, J.; Boo, K. A Study on Real-Time Detection Method of Lane and Vehicle for Lane Change Assistant System Using Vision System on Highway. Eng. Sci. Technol. Int. J. 2018, 21, 822–833. [Google Scholar] [CrossRef]
  3. Ye, Y.; Wang, Z.; Jones, D.; He, L.; Taylor, M.E.; Hollinger, G.A.; Zhang, Q. Bin-Dog: A Robotic Platform for Bin Management in Orchards. Robotics 2017, 6, 12. [Google Scholar] [CrossRef] [Green Version]
  4. González, R.; Pawlowski, A.; Rodríguez, C.; Guzmán, J.L.; Sánchez-Hermosilla, J. Design and Implementation of an Automatic Pressure-Control System for a Mobile Sprayer for Greenhouse Applications. Span. J. Agric. Res. 2012, 10, 939–949. [Google Scholar] [CrossRef] [Green Version]
  5. Wang, L.; Xiang, L.; Tang, L.; Jiang, H. A Convolutional Neural Network-Based Method for Corn Stand Counting in the Field. Sensors 2021, 21, 507. [Google Scholar] [CrossRef]
  6. Sampaio, G.S.; Silva, L.A.; Marengoni, M. 3D Reconstruction of Non-Rigid Plants and Sensor Data Fusion for Agriculture Phenotyping. Sensors 2021, 21, 4115. [Google Scholar] [CrossRef]
  7. Ochman, M. Hybrid approach to road detection in front of the vehicle. IFAC-PapersOnLine 2019, 52, 245–250. [Google Scholar] [CrossRef]
  8. Bietresato, M.; Carabin, G.; D’Auria, D.; Gallo, R.; Ristorto, G.; Mazzetto, F.; Vidoni, R.; Gasparetto, A.; Scalera, L. A tracked mobile robotic lab for monitoring the plants volume and health. In Proceedings of the 2016 12th IEEE/ASME International Conference on Mechatronic and Embedded Systems and Applications (MESA), Auckland, New Zealand, 29–31 August 2016; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  9. Ilesanmi, D.; Vincent, B.; Adefemi, A.; Bankole, O.; Johnson, K.P.; Khumbulani, M. Development and Performance Evaluation of a Robot for Lawn Mowing. Procedia Manuf. 2020, 49, 42–48. [Google Scholar] [CrossRef]
  10. Norris, W.R.; Patterson, A.E. System-Level Testing and Evaluation Plan for Field Robots: A Tutorial with Test Course Layouts. Robotics 2019, 8, 83. [Google Scholar] [CrossRef] [Green Version]
  11. Oliveira, L.F.P.; Moreira, A.P.; Silva, M.F. Advances in Agriculture Robotics: A State-of-the-Art Reviewand Challenges Ahead. Robotics 2021, 10, 52. [Google Scholar] [CrossRef]
  12. Aguiar, A.; dos Santos, F.; Cunha, J.; Sobreira, H.; Sousa, A. Localization and Mapping for Robots in Agriculture and Forestry: A Survey. Robotics 2020, 9, 97. [Google Scholar] [CrossRef]
  13. Hameeda, I.; la Cour-Harbo, A.; Osen, O. Side-to-side 3D coverage path planning approach for agricultural. Robot. Auton. Syst. 2013, 76, 36–45. [Google Scholar] [CrossRef]
  14. Lu, D.V.; Hershberger, D.; Smart, W.D. Layered costmaps for context-sensitive navigation. In Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 709–715. [Google Scholar] [CrossRef]
  15. Ball, D.; Upcroft, B.; Wyeth, G.; Corke, P.; English, A.; Ross, P.; Patten, T.; Fitch, R.; Sukkarieh, S.; Bate, A. Vision-based Obstacle Detection and Navigation for an Agricultural Robot. J. Field Robot. 2016, 33, 1107–1130. [Google Scholar] [CrossRef]
  16. Chen, Z.; Wang, S.; Wang, J.; Xu, K.; Lei, T.; Zhang, H.; Wang, X.; Liu, D.; Si, J. Control strategy of stable walking for a hexapod wheel-legged robot. ISA Trans. 2021, 108, 367–380. [Google Scholar] [CrossRef]
  17. Christiansen, P.; Nielsen, L.N.; Steen, K.A.; Jørgensen, R.N.; Karstoft, H. DeepAnomaly: Combining Background Subtraction and Deep Learning for Detecting Obstacles and Anomalies in an Agricultural Field. Sensors 2016, 16, 1904. [Google Scholar] [CrossRef] [Green Version]
  18. Aghi, D.; Mazzia, V.; Chiaberge, M. Local Motion Planner for Autonomous Navigation in Vineyards with a RGB-D Camera-Based Algorithm and Deep Learning Synergy. Machines 2020, 8, 27. [Google Scholar] [CrossRef]
  19. Kragh, M.F.; Christiansen, P.; Laursen, M.S.; Larsen, M.; Steen, K.A.; Green, O.; Karstoft, H.; Jørgensen, R.N. FieldSAFE: Dataset for Obstacle Detection in Agriculture. Sensors 2017, 17, 2579. [Google Scholar] [CrossRef] [Green Version]
  20. Christiansen, P.; Kragh, M.; Steen, K.A.; Karstoft, H.; Jørgensen, R.N. Platform for evaluating sensors and human detection in autonomous mowing operations. Precis. Agric. 2017, 18, 350–365. [Google Scholar] [CrossRef] [Green Version]
  21. Ross, P.; English, A.; Ball, D.; Upcroft, B.; Wyeth, G.; Corke, P. Novelty-based visual obstacle detection in agriculture. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 1699–1705. [Google Scholar] [CrossRef] [Green Version]
  22. Fleischmann, P.; Berns, K. A Stereo Vision Based Obstacle Detection System for Agricultural Applications. In Field and Service Robotics: Results of the 10th International Conference; Wettergreen, D.S., Barfoot, T.D., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 217–231. [Google Scholar] [CrossRef]
  23. Brown, J.; Sukkarieh, S. Improving Monocular Depth Prediction in Ambiguous Scenes Using a Single Range Measurement. IFAC-PapersOnLine 2019, 52, 355–360. [Google Scholar] [CrossRef]
  24. Steen, K.A.; Christiansen, P.; Karstoft, H.; Jørgensen, R.N. Using Deep Learning to Challenge Safety Standard for Highly Autonomous Machines in Agriculture. J. Imaging 2016, 2, 6. [Google Scholar] [CrossRef] [Green Version]
  25. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems—Volume 1 (NIPS’12), Lake Tahoe, NV, USA, 3–6 December 2012; Curran Associates Inc.: Red Hook, NY, USA, 2012; pp. 1097–1105. [Google Scholar]
  26. Kragh, M.; Jørgensen, R.N.; Pedersen, H. Object Detection and Terrain Classification in Agricultural Fields Using 3D Lidar Data. In Computer Vision Systems; Nalpantidis, L., Krüger, V., Eklundh, J.O., Gasteratos, A., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 188–197. [Google Scholar]
  27. Dvorak, J.S.; Stone, M.L.; Self, K.P. Object Detection for Agricultural and Construction Environments Using an Ultrasonic Sensor. J. Agric. Saf. Health 2016, 22, 107–119. [Google Scholar] [CrossRef]
  28. Vázquez-Arellano, M.; Griepentrog, H.W.; Reiser, D.; Paraforos, D.S. 3-D Imaging Systems for Agricultural Applications—A Review. Sensors 2016, 16, 618. [Google Scholar] [CrossRef] [Green Version]
  29. Pajares, G.; García-Santillán, I.; Campos, Y.; Montalvo, M.; Guerrero, J.M.; Emmi, L.; Romeo, J.; Guijarro, M.; Gonzalez-de Santos, P. Machine-Vision Systems Selection for Agricultural Vehicles: A Guide. J. Imaging 2016, 2, 34. [Google Scholar] [CrossRef] [Green Version]
  30. Wang, Q.; Meng, Z.; Liu, H. Review on Applicability of Vision-based Detection Technology in Field Obstacle Detection. IOP Conf. Ser. Earth Environ. Sci. 2020, 474, 032017. [Google Scholar] [CrossRef]
  31. Inoue, K.; Kaizu, Y.; Igarashi, S.; Imou, K. The development of autonomous navigation and obstacle avoidance for a robotic mower using machine vision technique. IFAC-PapersOnLine 2019, 52, 173–177. [Google Scholar] [CrossRef]
  32. Franzius, M.; Dunn, M.; Einecke, N.; Dirnberger, R. Embedded Robust Visual Obstacle Detection on Autonomous Lawn Mowers. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 June 2017. [Google Scholar] [CrossRef]
  33. Elfes, A. Using occupancy grids for mobile robot perception and navigation. Computer 1989, 22, 46–57. [Google Scholar] [CrossRef]
  34. Quigley, M.; Gerkey, B.; Conley, K.; Faust, J.; Foote, T.; Leibs, J.; Berger, E.; Wheeler, R.; Ng, A. ROS: An open-source Robot Operating System. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA) Workshop on Open Source Robotics, Kobe, Japan, 12–17 May 2009. [Google Scholar]
  35. Krata, D.; Ochman, M.; Panek, M.; Skoczeń, M.; Spyra, K.; Kulas, Z.; Sroczyński, D.; Pawłowski, A. Adaptive Smith Predictor Control Scheme for a Nonlinear Hydraulic System. In Proceedings of the 26th IEEE International Conference on Emerging Technologies and Factory Automation, ETFA, Vasteras, Sweden, 7–10 September 2021. [Google Scholar]
  36. Keselman, L.; Woodfill, J.I.; Grunnet-Jepsen, A.; Bhowmik, A. Intel(R) RealSense(TM) Stereoscopic Depth Cameras. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1267–1276. [Google Scholar] [CrossRef]
  37. Ochman, M.; Skoczeń, M.; Krata, D.; Panek, M.; Spyra, K.; Pawłowski, A. RGB-D Odometry for Autonomous Lawn Mowing. In Proceedings of the 20th International Conference on Artificial Intelligence and Soft Computing, ICAISC, Zakopane, Poland, 21–23 June 2021. [Google Scholar]
  38. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018. [Google Scholar]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:cs.CV/1512.03385. [Google Scholar]
  40. Rusu, R.B.; Cousins, S. 3D is here: Point Cloud Library (PCL). In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 1–4. [Google Scholar] [CrossRef] [Green Version]
  41. İyidir, İ.K.; Tek, F.B.; Kırcalı, D. Adaptive Visual Obstacle Detection for Mobile Robots Using Monocular Camera and Ultrasonic Sensor. In Proceedings of the Workshops and Demonstrations, Computer Vision—ECCV 2012, Florence, Italy, 7–13 October 2012; Fusiello, A., Murino, V., Cucchiara, R., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 526–535. [Google Scholar]
  42. John, V.; Mita, S. RVNet: Deep Sensor Fusion of Monocular Camera and Radar for Image-Based Obstacle Detection in Challenging Environments. In Image and Video Technology; Lee, C., Su, Z., Sugimoto, A., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 351–364. [Google Scholar]
  43. Li, Y.; Li, M.; Qi, J.; Zhou, D.; Zou, Z.; Liu, K. Detection of typical obstacles in orchards based on deep convolutional neural network. Comput. Electron. Agric. 2021, 181, 105932. [Google Scholar] [CrossRef]
  44. Powers, D. Evaluation: From Precision, Recall and F-Factor to ROC, Informedness, Markedness & Correlation. Mach. Learn. Technol. 2011, 2, 37–63. [Google Scholar]
  45. Le, H.; Das, P.; Mensink, T.; Karaoglu, S.; Gevers, T. EDEN: Multimodal Synthetic Dataset of Enclosed garDEN Scenes. In Proceedings of the IEEE/CVF Winter Conference of Applications on Computer Vision (WACV), Waikola, HI, USA, 5–9 January 2021. [Google Scholar]
  46. Sager, C.; Janiesch, C.; Zschech, P. A survey of image labelling for computer vision applications. J. Bus. Anal. 2021, 1–20. [Google Scholar] [CrossRef]
  47. Monica, R.; Aleotti, J.; Zillich, M.; Vincze, M. Multi-label Point Cloud Annotation by Selection of Sparse Control Points. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017. [Google Scholar] [CrossRef]
Figure 1. Obstacle detection system overview.
Figure 1. Obstacle detection system overview.
Sensors 21 05292 g001
Figure 2. Autonomous mower platform used in the research.
Figure 2. Autonomous mower platform used in the research.
Sensors 21 05292 g002
Figure 3. Examples of RGB images with corresponding depth maps and reference masks (from left to right) for each sequence: Boulevard 1, Boulevard 2, Park, and Backyard (from top to bottom).
Figure 3. Examples of RGB images with corresponding depth maps and reference masks (from left to right) for each sequence: Boulevard 1, Boulevard 2, Park, and Backyard (from top to bottom).
Sensors 21 05292 g003
Figure 4. Example of the point cloud reconstruction—segmentation of the RGB image, depth image, and resulting point cloud with labels.
Figure 4. Example of the point cloud reconstruction—segmentation of the RGB image, depth image, and resulting point cloud with labels.
Sensors 21 05292 g004
Figure 5. An example of the occupancy grid with a resolution of 1.0 m. The red color represents the map cells that were incorrectly marked as free space for robot operation (FN). Blue cells are the correctly detected obstacles (TP). The cell (2,1) contains an obstacle that was not detected, and the closest map cell to it that represents the obstacle is (1,4). Therefore, the d min - obst = ( 2 , 1 ) ( 1 , 4 ) 2 3.16 m.
Figure 5. An example of the occupancy grid with a resolution of 1.0 m. The red color represents the map cells that were incorrectly marked as free space for robot operation (FN). Blue cells are the correctly detected obstacles (TP). The cell (2,1) contains an obstacle that was not detected, and the closest map cell to it that represents the obstacle is (1,4). Therefore, the d min - obst = ( 2 , 1 ) ( 1 , 4 ) 2 3.16 m.
Sensors 21 05292 g005
Figure 6. Example of (a) manually labeled point cloud (red—grass, green—obstacle, blue—person), and (b) reference map from Boulevard 1 sequence (green—grass, blue—obstacle, yellow—person, purple—unknown).
Figure 6. Example of (a) manually labeled point cloud (red—grass, green—obstacle, blue—person), and (b) reference map from Boulevard 1 sequence (green—grass, blue—obstacle, yellow—person, purple—unknown).
Sensors 21 05292 g006
Figure 7. Example scene from the Park sequence: (a) presents the RGB image of the scene, (b) shows the labeled reference point cloud, and (c,d) are the reference map and the result of the obstacle detection pipeline, respectively. The maps’ resolution is 0.1 m, and the model grass/obstacle/person was used.
Figure 7. Example scene from the Park sequence: (a) presents the RGB image of the scene, (b) shows the labeled reference point cloud, and (c,d) are the reference map and the result of the obstacle detection pipeline, respectively. The maps’ resolution is 0.1 m, and the model grass/obstacle/person was used.
Sensors 21 05292 g007
Table 1. The synthetic dataset used for experiments contained the data from the images marked with the following numbers [45]: ‘0003’, ‘0005’, ‘0008’, ‘0018’, ‘0026’, ‘0029’, ‘0030’, ‘0040’, ‘0053’, ‘0068’, ‘0080’, ‘0160’, ‘0200’, ‘0214’, ‘0225’, ‘0257’, ‘0272’, ‘0313’, ‘0330’, ‘0347’.
Table 1. The synthetic dataset used for experiments contained the data from the images marked with the following numbers [45]: ‘0003’, ‘0005’, ‘0008’, ‘0018’, ‘0026’, ‘0029’, ‘0030’, ‘0040’, ‘0053’, ‘0068’, ‘0080’, ‘0160’, ‘0200’, ‘0214’, ‘0225’, ‘0257’, ‘0272’, ‘0313’, ‘0330’, ‘0347’.
Illumination ConditionsClearCloudyOvercastSunsetTwilight
Number of images sets10,07010,018951910,05510,100
Table 2. Segmentation results for grass/obstacle model under different illumination conditions.
Table 2. Segmentation results for grass/obstacle model under different illumination conditions.
Illumination ConditionsClassAccuracy [%]F1 [%]Precision [%]Recall [%]Specificity [%]
ClearOther98.0398.8899.0698.792.92
Grass91.6790.4592.9298.7
CloudyOther98.2299.099.1998.8193.45
Grass92.090.5893.4598.81
OvercastOther98.1398.9499.1598.7393.54
Grass92.0590.693.5498.73
SunsetOther97.8898.899.0298.5892.49
Grass90.989.3792.4998.58
TwilightOther98.0498.999.0798.7292.63
Grass91.3490.0992.6398.72
Table 3. Segmentation results for grass/obstacle model for synthetic dataset.
Table 3. Segmentation results for grass/obstacle model for synthetic dataset.
ClassAccuracy [%]F1 [%]Precision [%]Recall [%]Specificity [%]
Other98.0698.999.198.7193.0
Grass91.5990.2193.098.71
Table 4. Segmentation results for grass/obstacle model for real-world dataset.
Table 4. Segmentation results for grass/obstacle model for real-world dataset.
SequenceClassAccuracy [%]F1 [%]Precision [%]Recall [%]Specificity [%]
Boulevard 1Other98.4898.4899.6997.2999.69
Grass98.4997.3299.6997.29
Boulevard 2Other96.2592.5092.4292.5997.47
Grass97.5097.5397.4792.59
ParkOther87.3675.0673.1977.0290.75
Grass91.5492.3490.7577.02
BackyardOther98.6998.6498.4898.8098.59
Grass98.7398.8898.5998.80
Table 5. Segmentation results for grass/obstacle/person model for real-world dataset.
Table 5. Segmentation results for grass/obstacle/person model for real-world dataset.
SequenceClassAccuracy [%]F1 [%]Precision [%]Recall [%]Specificity [%]
Boulevard 1Other98.2298.0697.0799.0797.52
Grass98.6498.6899.9197.4899.90
Person99.4489.8384.6395.7299.54
Boulevard 2Other96.2892.4192.3392.4997.51
Grass96.6997.8097.8597.7493.54
Person99.050.000.0099.50
ParkOther97.3694.7592.5097.1197.44
Grass97.5098.3299.1297.5597.34
Person99.8322.3225.4219.8999.93
Table 6. Obstacle detection results based on map updates using grass/obstacle model for synthetic dataset and various occupancy grid resolutions.
Table 6. Obstacle detection results based on map updates using grass/obstacle model for synthetic dataset and various occupancy grid resolutions.
Map Resolution [m]ClassAccuracy [%]F1 [%]Precision [%]Recall [%]Specificity [%]
0.05Unknown100.0100.0100.0100.0100.0
Obstacle99.4594.592.0697.0899.57
Grass99.4581.989.7375.3299.85
0.1Unknown100.0100.0100.0100.0100.0
Obstacle99.0792.8588.7497.3599.18
Grass99.0772.6288.2561.6999.83
Table 7. Summary statistics of the d min - obst measure evaluated on grass/obstacle model for synthetic dataset and various occupancy grid resolutions.
Table 7. Summary statistics of the d min - obst measure evaluated on grass/obstacle model for synthetic dataset and various occupancy grid resolutions.
Map Resolution [m] avg ( d min - obst ) [m] var ( d min - obst ) [m]
0.050.160.02
0.10.170.03
Table 8. Obstacle detection results based on map updates using grass/obstacle/person model and 10 cm occupancy grid resolution.
Table 8. Obstacle detection results based on map updates using grass/obstacle/person model and 10 cm occupancy grid resolution.
SequenceClassAccuracy [%]F1 [%]Precision [%]Recall [%]Specificity [%]
Boulevard 1Unknown100.0100.0100.0100.0100.0
Obstacle100.0100.0100.0100.0100.0
Grass100.0100.0100.0100.0100.0
Boulevard 2Unknown99.3399.5299.06100.097.75
Obstacle97.9581.0481.8280.2998.97
Grass98.3396.5197.6695.3999.27
ParkUnknown99.0899.3199.3199.3198.60
Obstacle96.4250.2734.1195.5196.44
Grass95.6992.6998.3887.6399.34
BackyardUnknown100.0100.0100.0100.0100.0
Obstacle99.9161.9763.3760.2799.96
Grass99.9999.8799.9299.9299.96
Table 9. Obstacle detection results based on map updates using grass/obstacle/person model and 5 cm occupancy grid resolution.
Table 9. Obstacle detection results based on map updates using grass/obstacle/person model and 5 cm occupancy grid resolution.
SequenceClassAccuracy [%]F1 [%]Precision [%]Recall [%]Specificity [%]
Boulevard 1Unknown99.8599.9199.82100.0099.24
Obstacle98.7462.0299.4245.0699.99
Grass98.8196.6093.6599.7598.62
Boulevard 2Unknown99.6099.7599.5099.9997.93
Obstacle99.1684.4088.0781.0399.68
Grass99.3498.0098.5297.4999.71
ParkUnknown98.7399.1999.2299.1697.20
Obstacle97.7040.9826.1394.9997.72
Grass96.7391.7397.0586.9699.30
BackyardUnknown100.0100.0100.0100.0100.0
Obstacle99.9666.1269.9063.5799.99
Grass99.9799.9299.9799.9699.98
Table 10. Summary statistics of the d min - obst measure evaluated on grass/obstacle/person model for real world dataset in different image sequences and 10 cm occupancy grid resolutions.
Table 10. Summary statistics of the d min - obst measure evaluated on grass/obstacle/person model for real world dataset in different image sequences and 10 cm occupancy grid resolutions.
SequenceMap Resolution [m] avg ( d min - obst ) [m] var ( d min - obst ) [m]
Boulevard 10.10.130.002
0.050.10.003
Boulevard 20.10.380.18
0.050.310.16
Park0.12.745.5
0.053.135.34
Backyard0.10.140.01
0.050.130.07
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Skoczeń, M.; Ochman, M.; Spyra, K.; Nikodem, M.; Krata, D.; Panek, M.; Pawłowski, A. Obstacle Detection System for Agricultural Mobile Robot Application Using RGB-D Cameras. Sensors 2021, 21, 5292. https://doi.org/10.3390/s21165292

AMA Style

Skoczeń M, Ochman M, Spyra K, Nikodem M, Krata D, Panek M, Pawłowski A. Obstacle Detection System for Agricultural Mobile Robot Application Using RGB-D Cameras. Sensors. 2021; 21(16):5292. https://doi.org/10.3390/s21165292

Chicago/Turabian Style

Skoczeń, Magda, Marcin Ochman, Krystian Spyra, Maciej Nikodem, Damian Krata, Marcin Panek, and Andrzej Pawłowski. 2021. "Obstacle Detection System for Agricultural Mobile Robot Application Using RGB-D Cameras" Sensors 21, no. 16: 5292. https://doi.org/10.3390/s21165292

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop