Next Article in Journal
A Change of Theme: The Role of Generalization in Thematic Mapping
Next Article in Special Issue
Vehicle Detection in Very-High-Resolution Remote Sensing Images Based on an Anchor-Free Detection Model with a More Precise Foveal Area
Previous Article in Journal
Estimating and Interpreting Fine-Scale Gridded Population Using Random Forest Regression and Multisource Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Weighted Ensemble Object Detection with Optimized Coefficients for Remote Sensing Images

1
Department of Computer Engineering, Faculty of Technology, Gazi University, Ankara 06560, Turkey
2
Biomedical Engineering Department, Afyon Kocatepe University, Afyon 03300, Turkey
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2020, 9(6), 370; https://doi.org/10.3390/ijgi9060370
Submission received: 26 March 2020 / Revised: 20 April 2020 / Accepted: 29 May 2020 / Published: 4 June 2020
(This article belongs to the Special Issue Machine Learning for High Spatial Resolution Imagery)

Abstract

:
The detection of objects in very high-resolution (VHR) remote sensing images has become increasingly popular with the enhancement of remote sensing technologies. High-resolution images from aircrafts or satellites contain highly detailed and mixed backgrounds that decrease the success of object detection in remote sensing images. In this study, a model that performs weighted ensemble object detection using optimized coefficients is proposed. This model uses the outputs of three different object detection models trained on the same dataset. The model’s structure takes two or more object detection methods as its input and provides an output with an optimized coefficient-weighted ensemble. The Northwestern Polytechnical University Very High Resolution 10 (NWPU-VHR10) and Remote Sensing Object Detection (RSOD) datasets were used to measure the object detection success of the proposed model. Our experiments reveal that the proposed model improved the Mean Average Precision (mAP) performance by 0.78%–16.5% compared to stand-alone models and presents better mean average precision than other state-of-the-art methods (3.55% higher on the NWPU-VHR-10 dataset and 1.49% higher when using the RSOD dataset).

1. Introduction

At present, with the enhancement of remote sensing technologies, remote sensing image object detection has gained popularity thanks to its successful civil and military applications (e.g., urban monitoring, traffic monitoring, agricultural applications, and landscape planning). However, various difficulties remain in the field of remote sensing object detection, such as complicated and diverse views, the high costs of manual annotation, and differences in the instant detection of large scene images. Besides that, a typical feature of high-resolution images from planes or satellites is a highly detailed and mixed background. Object detection methods that use deep learning methods (which are increasing in popularity) have reached a state-of-art level [1].
Many studies on the detection of objects in remote sensing images are available in this field of study. For instance, Peicheng et al. [2] proposed a novel and effective approach to train a Rotation-Invariant Convolutional Neural Network (RICNN) model for advancing the performance of object detection, and Wang et al. [3] used a skip-connected encoder–decoder model to extract multiscale features from a full-size image. Wu et al. [4] detected remote sensing objects using Fourier-based rotation-invariant feature boosting (FRIFB). Dong et al. [5] used the Sigmoid Nonmaximum Suppression (Sig-NMS) method instead of the traditional nonmaximum suppression (NMS) to reduce the rate of undetectability for small targets. Gong et al. [6] introduced an object detection model that enriches feature representation and adopts the basic context information between objects. Cheng et al. [7] proposed a multiclass object detection feedback network (MODFN) using a top-down feedback mechanism based on a traditional feedforward network. Wang et al. [8] achieved object detection using a unified framework that can gather contextual information at multiple scales along with feature maps at the same scale. Zhang et al. [9] proposed a Double-Net model with multiple CNN channels, where each channel corresponds to a certain direction of rotation, and all CNNs share the same weights. Cheng et al. [10] developed a rotation-invariant framework based on the Collection of Part Detectors (COPD) for multiclass object detection. Li et al. [11] proposed a dual-channel feature fusion network capable of learning regional and contextual properties along two independent paths. Wang et al. [12] proposed a new region proposal network that is anchor-free and sliding-window-free to develop a two-stage object detection network. Körez et al. [13] improved the Faster R-CNN algorithm via the deformable convolution and the weight standardization techniques for remote sensing object detection using low-capacity Graphic Processing Units (GPUs). Qui et al. [14] proposed a novel end-to-end Adaptive Aspect Ratio Multiscale Network (A2RMNet), which is a multiscale feature gate fusion network used to adaptively integrate the multiscale features of objects. Xu et al. [15] proposed a CNN structure called deformable ConvNet, which adds offsets to the convolution layers to address geometric modeling in object recognition. Zhou et al. [16] proposed an encoder–encoder architecture called a rotated feature network (RFN), which produces rotation-sensitive feature maps (RS) for regression and rotation-invariant feature maps (RI) for classification. For a comprehensive and recent survey, see [17].
In recent years, studies using ensemble models have been prominent for object detection in remote sensing images [18,19,20,21,22,23]. In these studies, multiple multilayer perceptions, conditional random fields, or CNN ensembles were used to create structurally complete models. Then, the models were trained, and object detection was performed.
All existing models are based on specific structural conditions. Therefore, it is difficult to make changes to the models. This led researchers to focus on training ensemble models. However, training an ensemble model is not practical. However, in all of these studies, training was carried out on such ensemble models. A training-independent ensemble operation would be more practical and effective. In this study, we propose a flexible and training-independent weighted ensemble object model that uses three different object detection models (Single Shot Multibox Detector (SSD) [24], RetinaNet [25], and Improved Faster Regional Convolutional Neural Network (Faster R-CNN) [26]) that are independently trained on the same dataset. The SSD, RetinaNet, and Faster R-CNN models have achieved state-of-art success in object detection and are pioneers in the field. We chose the SSD, RetinaNet, and Improved Faster R-CNN models because they can optimize Faster R-CNN and previously achieved very successful results on the Northwestern Polytechnical University Very High Resolution 10 (NWPU-VHR10) and Remote Sensing Object Detection (RSOD) datasets. Our contributions are as follows:
  • A weighted ensemble object detection model using three different object detection models trained on the same dataset is proposed;
  • The proposed model is independent of the object detection models it uses as input. This means that two or more different types of object detection models can be used in the proposed model. Since our model has a weighted ensemble structure, the number of different structures of the models used as input will not affect the overall structure of the model;
  • Our study is the first to perform object detection of remote sensing images using multiple object detection models as coefficient-weighted ensembles with optimized coefficients that are independently trained on the same dataset.
In the second section of this study, the materials and methods we used are described. In the third section, the proposed optimized coefficient-weighted ensemble model is outlined. The fourth section focuses on the test environment, evaluation criteria, and explanation of the datasets (NWPU-VHR10 and RSOD) that we used to measure the performance of our proposed model, as well as experiments using different coefficients (including optimal coefficients) and a comparison of our model with other studies using the same datasets. The conclusions of the study are summarized in Section 5, and Section 6 discusses future work.

2. Methods

In this section, the structures and concepts used by the remote sensing object detection model introduced in our study are explained.

2.1. Single Shot Multibox Detector (SSD)

The SSD divides the output area of the bounding boxes into a series of default boxes with different aspect ratios and scales for each feature map location [24]. While making predictions, the algorithm generates points for the presence of each category of objects in the default box and adjusts the sizes of the boxes to better match the shape of the object. In addition, predictions from multiple attribute maps with different resolutions are combined so that objects of different sizes can be handled easily (Figure 1).

2.2. RetinaNet

RetinaNet is a single unified network consisting of a backbone network and two task-specific subnets. To increase the object detection success, RetinaNet uses Focal Loss [25], which is a new type of loss that reduces the relative loss of well-classified samples and pays more attention to difficult-to-detect and misclassified samples. In RetinaNet, the backbone is responsible for calculating a convolutional property map over the entire input image. The first subnet classifies the backbone output, and the second subnet performs the convolution, thereby limiting box regression (Figure 2).

2.3. Improved Faster R-CNN

The present model is based on the latest Faster R-CNN (Figure 3) [26], a state-of-the-art object detection system. The deformable convolution [27] technique is used to overcome the weaknesses of the regular convolution structure used in the Faster R-CNN model for detecting small and mixed objects in remote sensing images [13]. With the Feature Pyramid Network (FPN) [28] technique, the high-resolution features in the shallow layers of the remote sensing images are transferred to the network. The Weight Standardization (WS) [29] technique, which reduces the batch size to provide deep learning training without performance problems on low power / single GPU systems, such as those using a single GPU, was also added to the model.

2.4. General Concepts of Object Detection

In this subsection, the concepts we use within the scope of object detection are briefly defined. Those are;
  • Bounding Boxes: These are the rectangular boxes drawn around the objects to make the detected objects more prominent after the object detection is complete;
  • Common Box: This is the box used to express the bounding boxes from three different object detection models, which are used as the input for the Weighted Ensemble Block; a single bounding box is called common box;
  • Confidence Score: This is the probability that an anchor box contains an object and is usually predicted by a classifier. The confidence score contains cut-off predictions [26,27,28].

3. Proposed Model

In this section, we describe the ensemble algorithm that we used to combine the included object detection models, the optimization of the coefficients, and the outputs of these object detection models. Figure 4 shows a block diagram of the object detection model that we propose in our study.

3.1. Object Detection Models

In the proposed model, three different object detection models trained on the same dataset are used. These models are Single Shot Multibox Detector (SSD), RetinaNet, and Improved Faster R-CNN for remote sensing images. Improved Faster R-CNN is a two-stage object detection model, while SSD and RetinaNet are single-stage object detection models. These object detection models are independently trained. Details of the training process are described below:
  • Twenty-four Epoch training was performed, and the batch size was 64;
  • Learning rate = 0.0025 and momentum = 0.9. A gradual learning rate technique was applied by reducing the learning rate. This was done by dividing 10 by the 16th epoch and 22nd epoch;
  • Batch Normalization was used for the SSD and RetinaNet models, and Weight Standardization was used for the Improved Faster R-CNN model;
  • Transfer learning was utilized for the trained Resnet50 model.

3.2. Weighted Ensemble Block

This algorithm corresponds to the Weighted Ensemble Block in Figure 4. It takes three different object detection results as the input from this algorithm and produces a single object detection result. The object detection concept involves the task of determining the position and category of multiple objects in an image. An object detection model is a function that, when given an image, returns a list of detections (D = [d1, …, dN]), where each di is given by a triple [bi, ci, si] that consists of a bounding box, bi, the corresponding category, ci, and the corresponding confidence score, si.
The input of our ensemble algorithm is a List of Detections (LD) for a given image using three different object detection models (LD = [D1, D2, D3]). Each LD has different bi, ci, and si values. The algorithm steps are as follows:
(1) The detections in LD are grouped according to the class values and the overlap of the bounding boxes. Taking the two bounding boxes as b1 and b2, the following IoU equation is used to find the overlapped region between them.
I o U ( b 1 ,   b 2 ) = a r e a   (   b 1     b 2   ) a r e a   ( b 1     b 2   )
(2) An IoU threshold value of 0.5 is frequently used in related object detection studies, including the studies mentioned in the introduction. Thus, bounding boxes with IoU > 0.5 are assigned to a new Output List (LO). The LO has bi and Si values for each detection. It is necessary to recalculate a common box (X1, Y1, X2, Y2) and confidence score (S) using the weight coefficients (ki) for the multiple different boxes in the LO clusters. These calculations are made with the following equations:
S = k 1   S 1 + k 2   S 2 + k 3   S 3 k 1 + k 2 + k 3
X 1 = k 1   X 1 1 + k 2   X 1 2 + k 3   X 1 3 k 1 + k 2 + k 3
X 2 = k 1   X 2 1 + k 2   X 2 2 + k 3   X 2 3 k 1 + k 2 + k 3
Y 1 = k 1   Y 1 1 + k 2   Y 1 2 + k 3   Y 1 3 k 1 + k 2 + k 3
Y 2 = k 1   Y 2 1 + k 2   Y 2 2 + k 3   Y 2 3 k 1 + k 2 + k 3 .
When calculating the confidence score in Equation (2), the weight coefficients and multiplication results of all confidence scores are averaged. In the same way, the coordinates of the bounding box are calculated by multiplying the individual weight coefficients in each of them (Equations (3)–(6)).
(3) If least two of the detections in the LD overlap, the above operations are performed. The resulting ensemble detections are added to the Predictions List (PL). At this point, some detections may not overlap at all. Those with a confidence score of (c)> 0.5 are also assigned to the PL.
(4) The detections in the PL are considered the final predictions, and the weighted ensemble process ends.

3.3. Optimizing Coefficients

This is the process for calculating the optimal coefficients according to the detection results obtained from the three models we used. In this process, the coefficients that converge the X1, Y1, X2, and Y2 coordinate values of the three models closest to the X1, Y1, X2, and Y2 coordinate values of the labels in the test dataset are calculated via nonlinear programming [30]. In this process, optimization is achieved by considering all test data as a whole. The calculations are as follows:
Y ^ i = n = 1 3 k n   X i n k s u m   ,   X i n = ( X 1 n   ,   Y 1 n   ,   X 2 n   ,   Y 2 n )
k s u m = k 1 + k 2 + k 3
C o n s t r a i n t s   :   n = 1 3 k n = 1     ,   k n   0
Z m i n = i = 1 K (   Y i   Y ^ i   ) 2 ) .
In Equation (8), Y ^ i represents the sum of the estimates, k n denotes the coefficients, and X i n refers to the coordinate values (X1, Y1, X2, Y2) of the ith prediction of the nth model. Equation (9) is the formula for the sum of coefficients, and Equation (10) is the optimization objective function that minimizes the sum of the squares of the difference between the real coordinate values and the estimated coordinate values. The constraints of this optimization objective function are given in Equation (9). For the optimization objective function to find the minimum value, the Mean Squared Deviation (MSD) [31] technique is used. The values of k 1 ,   k 2 ,   k 3 that provide the Z m i n value are considered the optimal coefficients.

4. Results and Discussion

An experimental setup environment was created to evaluate the performance of the proposed model on the NWU-VHR10 and RSOD datasets. Three object detection models used as the input for the ensemble model and ensemble object detection results under different combinations of coefficients were compared. Then, a comparison was carried out with other related studies using the NWPU-VRH10 and RSOD datasets to evaluate how successful the proposed model was.

4.1. Environment and Evaluation Criteria

For the physical environment, the PC used for the experiments features an Intel® Core ™ i5 2.4 GHz CPU, with 6 GB RAM, a Geforce GTX 1080 graphics card, and the Ubuntu 16.04 LTS operating system. The proposed model was written in the python programming language using the pytorch deep learning library. To evaluate the performance of the proposed model, the precision–recall curve (PRC) and mean average precision (mAP) [32] criteria, which are used to measure the success of many object detection models [33,34,35,36,37,38,39] in this area, are used.
  • Precision–Recall Curve (PRC): Precision is used to indicate the accuracy of the detection rate of true positive values, and recall is used to indicate the ratio of true positive values detected correctly [40]. The number of true positives is TP, the number of false positives is FP, and the number of false negatives is FN. The precision–recall curve value is calculated with the following formulas:
P r e c i s i o n = T P ( T P + F P )
R e c a l l = T P ( T P + F N ) )
where the TP value used in Equations (11) and (12) refers to the situations where the area between the box overlaps, and the Ground Truth (GT) box is greater than 0.5 as a result of the detection. The opposite case is considered to be the FP. Further, if multiple detections coincide with the same GT box, only one is considered the TP, and the others are FP.
  • Mean Average Precision (mAP): This can be explained as the average of the precision values of all classes in the area under the PRC. Therefore, the performance of an object detection model is measured by the height of the mAP value. A higher mAP value indicates better performance.

4.2. Dataset Preparation Phase

In the experiments, the NWPU-VHR10 dataset [41] and RSOD dataset [42] were used. These datasets are described below.
(1) NWPU-VHR10 dataset: As its name suggests, this dataset has 10 classes (aircraft, ship, storage tank, baseball diamond, tennis court, basketball court, highway court, port, bridge, and vehicle). Moreover, this dataset contains 650 positive images and 150 negative images (total 800 images), whose spatial resolution ranges from 0.5 to 2 m.
In the experiments, positive images in the dataset were used in two different ways (60% training—390 images, 20% verification—130 images, 20% testing—130 images, and 50% training—325 images, 20% verification—130 images, 30% testing—195 images). The set of 650 images containing objects consists of 565 images from Google Earth with a spatial resolution ranging from 0.2 m to 2 m and 85 sharpened images with a spatial resolution of 0.08 m from the Vaihingen dataset [43].
(2) RSOD dataset: The RSOD dataset contains 976 images downloaded from Google Earth and Tianditu [44]; the spatial resolutions of these images range from 0.3 to 3 m. This dataset consists of 6950 total object instances covered by four object classes, including 1586 oil tanks, 4993 airplanes, 180 overpasses, and 191 playgrounds. In the experiments, images in this dataset were used in two different ways, as in the NWU-VHR10 dataset (60% training—586 images, 20% verification—195 images, 20% testing—195 images, and 50% training—488 images, 20% verification—195 images, 30% testing—293 images).
The training objects in the NWU-VHR10 and RSOD datasets are few in number, which directly affects the success of the model. Data augmentation [45] was used to eliminate this disadvantage. The operations applied in the data augmentation process to the images in the datasets are presented sequentially in Figure 5.

4.3. Results with Different Coefficient Combinations and Optimal Coefficient Values

For the NWPU-VHR10 and RSOD datasets, the three object detection models used as the input for the ensemble model and the ensemble object detection results made with the different combinations and optimal values of the coefficients were compared. To better evaluate the performance of the proposed model, the comparison process was done according to two different train–test ratios frequently used in the literature. A comparison of the results is given in Table 1 and Table 2.
In Table 1 and Table 2, k1, k2, and k3 are the coefficient values for SSD, RetinaNet, and the Improved Faster R-CNN models, respectively.
The most successful results were obtained from the experiments using the optimal coefficients. Further, all ensemble coefficient combinations achieved high success in the single object detection models.

4.4. Comparing Proposed Model with Other Studies

In this section, we compare the proposed model to other studies that use both the NWPU-VHR10 dataset and the RSOD dataset.
(1) Comparision of the NWPU-VHR10 dataset: A comparison with other studies that used the NWPU-VRH10 dataset in this field was carried out to evaluate how successful the proposed model was. The comparison results are given in Table 3. The most successful results per class are marked with underlined bold font in the table.
(2) Comparision of the RSOD dataset: A comparison was made with other studies in this field that used the RSOD dataset to assess how successful our proposed model was, like with the NWPU-VHR10 dataset. Existing studies using the RSOD dataset are less common than those using the NWPU-VHR 10 dataset. The comparison results are given in Table 4. The most successful results per class are marked with underlined bold font in the table.
Using our weighted ensemble object detection model, we performed object detection for all classes in the test dataset. The object detection results for the test images of the proposed model are provided in Appendix A. In Appendix A, true positive detections are marked in green, false-positive detections are marked in red, and false negative detections are marked with a blue box.

4.5. Discussion

In this study, three parameters that can affect the object detection performance of the model relative to the dataset were considered. These parameters are the number of images containing the object in the dataset, the number of objects in each image, and the size of the object in the image. As noted in other studies [2,5,8,10,16] (using the comparisons in Table 3 and Table 4), the numbers of positive image sets from the NWPU-VHR10 dataset (650 images) and the images of the RSOD dataset (976 images) were too few for deep learning object detection models. We aimed to solve this problem with data augmentation. However, data augmentation alone was not enough to eliminate class imbalance. To eliminate the imbalance between classes, we combined the strengths of these models by using coefficients rather than designing a separate loss function for these three object detection models. Each model was trained independently using its own loss function. Then, the optimal coefficients were calculated using the MSD technique. In this way, the proposed model detects objects using the optimal coefficient values. In our experiments, we observed the effects of these coefficient changes on success.
As shown in Table 1 and Table 2, the optimal ensemble coefficients achieved greater success than the standalone object detection models that we used as input for the proposed model. In addition, other ensemble coefficient combinations achieved excellent success using single object detection models. In the conducted experiments, the proposed model increased the performance of classification for the NWPU-VHR10 and RSOD datasets by 11% to 16.5% for SSD, 0.78% to 6% for RetinaNet, and 0.88% to 4.54% for Improved Faster R-CNN. As shown by the experimental results, the highest improvement in classification performance was achieved with the SSD model because our model eliminated the disadvantages of the SSD model by applying the RetinaNet and Improved Faster R-CNN models, which are more successful than the SSD model.
As mentioned earlier, the small number of images found in the NWPU-VHR10 and RSOD datasets, which are frequently used in remote sensing literature, is too low for deep learning models. Thus, while greater success would be expected by training the model with more data under normal conditions, in our experiments with 80% training data, lower values were obtained than in experiments with 70% training data. This situation can be explained in two ways:
  • Overfitting: The proposed model experienced some overfitting in experiments conducted at a rate of 80% training, despite the use of data augmentation;
  • The Size of the Test Data: The object detection success of the model could be better measured by testing with more data. Accordingly, better results were obtained in tests with 30% testing data.
The results given in Table 3 reveal that the proposed model provides better results than those in other studies. Although our method ensures high performance, we observed a discrepancy in the detection accuracy for the same category of object (e.g., storage tank) as in the DODN [12] and Improved Faster R-CNN [13] models. This is mainly because of the imbalance among the classes in the NWPU-VHR10 database. The object detection rate of the classes (e.g., planes) with a large training dataset is high, while the object detection success of the classes (e.g., storage tank) with a small training dataset is quite low. Although data augmentation improved the training performance, its contribution remains limited. Table 4 shows that the proposed model did not experience this problem with the RSOD dataset. This is because the amount of image data used to train deep learning models directly affects the object detection success of the model (the RSOD dataset contains 50% more images (325 images) than the NWPU-VRH10 dataset).
Despite the inadequacies of this dataset, the proposed model achieved high success in detecting objects of different orientations and sizes in the test images. Moreover, our model resulted in higher performance results than those in other studies using the same dataset for object detection. The proposed model achieved a higher mAP than other studies in six of the ten object classes in the NWPU-VHR10 dataset, while in the RSOD dataset, our model achieved higher mAP success in two of the four object classes. We believe that the success of the proposed model is the result of the following:
  • Our model combines the powerful features of three different models (SSD, RetinaNet, and Improved Faster R-CNN), which have proven themselves successful for object detection;
  • We calculated the optimal coefficients of the object detection models using the MSD technique;
  • Each object detection model is subjected to an independent training process, so there is no overfitting problem.

5. Conclusions

In this study, an object detection model that performs the optimized coefficient-weighted ensemble operations for the outputs of three different object detection models trained on the same dataset was proposed. The proposed model was designed to provide an output with an optimized coefficient-weighted ensemble by taking two or more object detection methods as its input. In addition to providing an ensemble output by multiplying the most successful determinations of the object detection models used by the input with the coefficients determined through optimization using the MSD technique, the results obtained under various coefficient changes were more successful than the results with the object detection models used as input. The object detection performance of our proposed model was assessed using the NWPU-VHR10 and RSOD datasets. The highlights of our model include combining the strengths of SSD, RetinaNet, and Improved Faster R-CNN models as inputs and training these models separately to prevent overfitting. Therefore, the results of the experiments reveal that our model can obtain better results than current models that use the same datasets. In summary, no matter how different the models used as inputs and the hyperparameters of these models are, our ensemble model, which can use the strengths of the input models most effectively through optimal coefficients, achieves better results than the models we used as inputs.

6. Future Work

In our future work, studies will be continued on the following topics:
  • Different models and their experimental results will be analyzed by multiplying these models with different coefficients to develop a dataset to measure the success of the coefficients used for the model. Then, a simple neural network will be trained with this dataset. Using this training, we will attempt to reveal the predicted success of the network for the purpose of determining the most successful weighted ensemble method with machine learning. We will also test how successful the final results of the machine learning are through experiments using the proposed model;
  • To increase the object detection success of our model, a future study will be carried out focusing on reducing the effects of the imbalance between classes by designing the loss functions of the object detection methods (SSD, RetinaNet, and the Improved Faster R-CNN) as a single loss function.

Author Contributions

Conceptualization, Atakan Körez, Necaattin Barışçı, Aydın Çetin, and Uçman Ergün; data curation, Atakan Körez, Necaattin Barışçı, Aydın Çetin, and Uçman Ergün; formal analysis, Atakan Körez, Necaattin Barışçı, Aydın Çetin, and Uçman Ergün; investigation, Atakan Körez, Necaattin Barışçı, Aydın Çetin, and Uçman Ergün; methodology, Atakan Körez, Necaattin Barışçı, Aydın Çetin, and Uçman Ergün; project administration, Necaattin Barışçı and Aydın Çetin; resources, Atakan Körez, Necaattin Barışçı, Aydın Çetin, and Uçman Ergün; software, Atakan Körez; supervision, Necaattin Barışçı; validation, Atakan Körez, Necaattin Barışçı, Aydın Çetin, and Uçman Ergün; visualization, Atakan Körez and Uçman Ergün; writing—original draft, Atakan Körez, and Necaattin Barışçı; writing—review and editing, Atakan Körez, Necaattin Barışçı, and Aydın Çetin. All authors have read and agreed to the published version of the manuscript.

Funding

There was no external funding.

Acknowledgments

We would like to thank TÜBİTAK ULAKBİM High Performance and Grid Calculation Center for the high calculation servers we have tested the model created within the scope of the study. The authors thank all the reviewers for their comments, which improved the quality of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Object Detection Results of the Weighted Ensemble Object Detection Model

Figure A1. Object detection examples on the NWPU-VHR10 dataset: (a) plane; (b) baseball diamond; (c) vehicle.
Figure A1. Object detection examples on the NWPU-VHR10 dataset: (a) plane; (b) baseball diamond; (c) vehicle.
Ijgi 09 00370 g0a1
Figure A2. Object detection examples on an RSOD dataset: (a) aircraft; (b) oil tank; (c) playground.
Figure A2. Object detection examples on an RSOD dataset: (a) aircraft; (b) oil tank; (c) playground.
Ijgi 09 00370 g0a2

References

  1. Jiao, L.; Zhang, F.; Liu, F.; Yang, S.; Li, L.; Feng, Z.; Qu, R. A Survey of Deep Learning-Based Object Detection. IEEE Access 2019, 7, 128837–128868. [Google Scholar] [CrossRef]
  2. Peicheng, Z.; Cheng, G.; Junwei, H. Learning Rotation-Invariant Convolutional Neural Networks for Object Detection in VHR Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7416. [Google Scholar]
  3. Wang, C.; Bai, X.; Wang, S.; Zhou, J.; Ren, P. Multiscale Visual Attention Networks for Object Detection in VHR Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2019, 16, 309–314. [Google Scholar] [CrossRef]
  4. Wu, X.; Hong, D.; Chanussot, J.; Xu, Y.; Tao, R.; Wang, Y. Fourier-Based Rotation-Invariant Feature Boosting: An Efficient Framework for Geospatial Object Detection. IEEE Geosci. Remote Sens. Lett. 2019, 1–5. [Google Scholar] [CrossRef] [Green Version]
  5. Dong, R.; Xu, D.; Zhao, J.; Jiao, L.; An, J. Sig-NMS-Based Faster R-CNN Combining Transfer Learning for Small Target Detection in VHR Optical Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8534–8545. [Google Scholar] [CrossRef]
  6. Gong, Y.; Xiao, Z.; Tan, X.; Sui, H.; Xu, C.; Duan, H.; Li, D. Context-Aware Convolutional Neural Network for Object Detection in VHR Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2019, 58, 34–44. [Google Scholar] [CrossRef]
  7. Cheng, B.; Zheng-zhou, L.; Qingqing, W.; Bo, L.; Huihui, Y.; Lin, Q.; Bo, Q. Multi-Class Objects Detection Method in Remote Sensing Image Based on Direct Feedback Control for Convolutional Neural Network. IEEE Access 2019, 7, 144691–144709. [Google Scholar] [CrossRef]
  8. Wang, P.; Sun, X.; Diao, W.; Fu, K. FMSSD: Feature-Merged Single-Shot Detection for Multiscale Objects in Large-Scale Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2019, 38, 3377–3390. [Google Scholar] [CrossRef]
  9. Zhang, Z.; Jiang, R.; Mei, S.; Zhang, S.; Zhang, Y. Rotation-Invariant Feature Learning for Object Detection in VHR Optical Remote Sensing Images by Double-Net. IEEE Access 2019, 8, 20818–20827. [Google Scholar] [CrossRef]
  10. Cheng, G.; Han, J.; Zhou, P.; Guo, L. Multi-class geospatial object detection and geographic image classification based on collection of part detectors. ISPRS J. Photogramm. Remote Sens. 2014, 98, 119–132. [Google Scholar] [CrossRef]
  11. Li, K.; Cheng, G.; Bu, S.; You, X. Rotation-Insensitive and Context-Augmented Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2337–2349. [Google Scholar] [CrossRef]
  12. Wang, C.; Shi, J.; Yang, X.; Zhou, Y.; Wei, S.; Li, L.; Zhang, X. Geospatial Object Detection via Deconvolutional Region Proposal Network. IEEE J. Sel. Top. Appl. Obs. Remote Sens. 2019, 12, 1–14. [Google Scholar] [CrossRef]
  13. Körez, A.; Barışçı, N. Object Detection with Low Capacity GPU Systems Using Improved Faster R-CNN. Appl. Sci. 2019, 10, 83. [Google Scholar] [CrossRef] [Green Version]
  14. Qiu, H.; Li, H.; Wu, Q.; Meng, F.; Ngan, K.N.; Shi, H. A2RMNet: Adaptively Aspect Ratio Multi-Scale Network for Object Detection in Remote Sensing Images. Remote Sens. 2019, 11, 1594. [Google Scholar] [CrossRef] [Green Version]
  15. Xu, Z.; Xu, X.; Wang, L.; Yang, R.; Pu, F. Deformable ConvNet with Aspect Ratio Constrained NMS for Object Detection in Remote Sensing Imagery. Remote Sens. 2017, 9, 1312. [Google Scholar] [CrossRef] [Green Version]
  16. Zhou, K.; Zhang, Z.; Gao, C.; Liu, J. Rotated Feature Network for Multiorientation Object Detection of Remote-Sensing Images. IEEE Geosci. Remote Sens. Lett. 2020, 1–5. [Google Scholar] [CrossRef]
  17. Cheng, G.; Han, J. A Survey on Object Detection in Optical Remote Sensing Images. ISPRS J. Photogramm. Remote Sens. 2016, 117, 11–28. [Google Scholar] [CrossRef] [Green Version]
  18. Zhong, P.; Wang, R. A Multiple Conditional Random Fields Ensemble Model for Urban Area Detection in Remote Sensing Optical Images. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3978–3988. [Google Scholar] [CrossRef]
  19. Gokaraju, B.; Durbha, S.; King, R.; Younan, N.H. Ensemble Methodology Using Multistage Learning for Improved Detection of Harmful Algal Blooms. IEEE Geosci. Remote Sens. Lett. 2012, 9, 827–831. [Google Scholar] [CrossRef]
  20. Roy, M.; Routaray, D.; Ghosh, S.; Ghosh, A. Ensemble of Multilayer Perceptrons for Change Detection in Remotely Sensed Images. IEEE Geosci. Remote Sens. Lett. 2014, 11, 49–53. [Google Scholar] [CrossRef]
  21. Yu, Y.; Liu, F. Aerial Scene Classification via Multilevel Fusion Based on Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2018, 15, 287–291. [Google Scholar] [CrossRef]
  22. Dede, M.; Aptoula, E.; Genc, Y. Deep Network Ensembles for Aerial Scene Classification. IEEE Geosci. Remote Sens. Lett. 2019, 16, 732–735. [Google Scholar] [CrossRef]
  23. Jiang, H.; Qiu, X.; Chen, J.; Liu, X.; Miao, X.; Zhuang, S. Insulator Fault Detection in Aerial Images Based on Ensemble Learning with Multi-level Perception. IEEE Access 2019, 7, 61797–61810. [Google Scholar] [CrossRef]
  24. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.; Berg, A. SSD: Single Shot MultiBox Detector. Eur. Conf. Comput. Vis. (ECCV) 2016, 21–37. [Google Scholar] [CrossRef] [Green Version]
  25. Lin, T.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal loss for object detection. IEEE Int. Conf. Comput. Vis. 2017, 42, 318–327. [Google Scholar]
  26. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. Proc. Adv. Neural Inform. Process. Syst. 2015, 39, 91–99. [Google Scholar] [CrossRef] [Green Version]
  27. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  28. Tsung-Yi, L.; Dollar, P.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  29. Qiao, S.; Wang, H.; Liu, C.; Shen, W.; Yuille, A. Weight Standardization. Arxiv Prepr. 2019, arXiv:1903.10520. [Google Scholar]
  30. Ruszczyński, A. Nonlinear Optimization; Princeton University Press: New York, NY, USA, 2006; pp. 454–480. [Google Scholar]
  31. Lehmann, E.L.; Casella, G. Theory of Point Estimation, 2nd ed.; Springer Texts in Statics: New York, NY, USA, 1998. [Google Scholar]
  32. Liu, L.; Özsu, M.T. Encyclopedia of Database Systems; Springer: Boston, MA, USA, 2009. [Google Scholar]
  33. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A Large-Scale Hierarchical Image Database. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. 2009, 165–171. [Google Scholar] [CrossRef] [Green Version]
  34. Cheng, G.; Han, J.; Li, K.; Qian, X.; Zhou, P.; Yao, X.; Hu, X. Object Detection in Remote Sensing Imagery Using a Discriminatively Trained Mixture Model. ISPRS J. Photogramm. Remote Sens. 2019, 85, 32–43. [Google Scholar] [CrossRef]
  35. Liu, W.; Ma, L.; Wang, J.; Chen, H. Detection of Multiclass Objects in Optical Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2019, 16, 791–795. [Google Scholar] [CrossRef]
  36. Pang, J.; Li, C.; Shi, J.; Xu, Z.; Feng, H. R2 -CNN: Fast Tiny Object Detection in Large-Scale Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5512–5524. [Google Scholar] [CrossRef] [Green Version]
  37. Hu, Y.; Li, X.; Zhou, N.; Yang, L.; Peng, L.; Xiao, S. A Sample Update-Based Convolutional Neural Network Framework for Object Detection in Large-Area Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2019, 16, 947–951. [Google Scholar] [CrossRef]
  38. Yang, Z.; Liu, Y.; Liu, L.; Tang, X.; Xie, J.; Gao, X. Detecting Small Objects in Urban Settings Using SlimNet Model. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8445–8457. [Google Scholar] [CrossRef]
  39. Lu, X.; Zhang, Y.; Yuan, Y.; Feng, Y. Gated and Axis-Concentrated Localization Network for Remote Sensing Object Detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 179–192. [Google Scholar] [CrossRef]
  40. Powers, D. Evaluation: From Precision, Recall and F-Measure to ROC, Informedness, Markedness and Correlation. J. Mach. Learn. Technol. 2011, 2, 37–63. [Google Scholar]
  41. NWPU-VHR10 Dataset. Available online: http://www.escience.cn/people/gongcheng/NWPU-VHR-10.html (accessed on 15 July 2019).
  42. Xiao, Z.; Liu, Q.; Tang, Z.; Zhai, X. Elliptic Fourier transformation-based histograms of oriented gradients for rotationally invariant object detection in remote-sensing images. Int. J. Remote Sens. 2015, 36, 618–644. [Google Scholar] [CrossRef]
  43. Vaihingen Dataset. Available online: http://www2.isprs.org/commissions/comm3/wg4/2d-sem-label-vaihingen.html (accessed on 7 January 2020).
  44. China Official Web Mapping Service. Available online: www.tianditu.gov.cn (accessed on 7 March 2020).
  45. Buslaev, A.; Parinov, A.; Khvedchenya, E.; Iglovikov, V.; Kalinin, A. Albumentations: Fast and Flexible Image Augmentations. Information 2020, 11, 125. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Block diagram of the Single Shot Multibox Detector (SSD).
Figure 1. Block diagram of the Single Shot Multibox Detector (SSD).
Ijgi 09 00370 g001
Figure 2. Architecture of RetinaNet: (a) first subnet and (b) second subnet.
Figure 2. Architecture of RetinaNet: (a) first subnet and (b) second subnet.
Ijgi 09 00370 g002
Figure 3. Network structure of the Improved Faster R-CNN.
Figure 3. Network structure of the Improved Faster R-CNN.
Ijgi 09 00370 g003
Figure 4. Block diagram of weighted ensemble object detection method with optimized coefficients.
Figure 4. Block diagram of weighted ensemble object detection method with optimized coefficients.
Ijgi 09 00370 g004
Figure 5. Data augmentation examples: (a) original picture; (b) blur; (c) vertical rotation; (d) horizontal rotation; (e) gamma conversion; (f) random image brightness.
Figure 5. Data augmentation examples: (a) original picture; (b) blur; (c) vertical rotation; (d) horizontal rotation; (e) gamma conversion; (f) random image brightness.
Ijgi 09 00370 g005
Table 1. Comparison of the results of three different object detection models and different coefficient combinations for the Northwestern Polytechnical University Very High Resolution 10 (NWPU-VHR10) dataset.
Table 1. Comparison of the results of three different object detection models and different coefficient combinations for the Northwestern Polytechnical University Very High Resolution 10 (NWPU-VHR10) dataset.
ModelCoefficientsmAP
(70% Training—30% Testing)
mAP
(80% Training—20% Testing)
SSD- 0.8320.790
Imp. Faster R-CNN- 0.9160.889
RetinaNet- 0.9200.881
Experiment 1k1= k2 = k3 = 1 0.9310.913
Experiment 2k1= 2, k2 = k3 = 1 0.9240.909
Experiment 3k1= k3 = 1, k2 = 2 0.9330.901
Experiment 4k1= k2 = 1, k3 = 2 0.9300.905
Experiment 5k1= k2 = 2, k3 = 1 0.9290.908
Experiment 6k1= k3 = 2, k2 = 1 0.9320.907
Experiment 7k2= k3 = 2, k1 = 1 0.9270.906
Experiments 8*k1= 0.0226, k2 = 0.6527, k3 = 0.3246 0.942 -
Experiments 9**k1= 0, k2 = 0.7982, k3 = 0.2017 -0.921
* Optimal Coefficients for 70% Training—30% Testing ** Optimal Coefficients for 80% Training—20% Testing.
Table 2. Comparison of the results of three different object detection models and different coefficient combinations for the Remote Sensing Object Detection (RSOD) dataset.
Table 2. Comparison of the results of three different object detection models and different coefficient combinations for the Remote Sensing Object Detection (RSOD) dataset.
ModelCoefficientsmAP
(70% Training—30% Test)
mAP
(80% Training—20% Testing)
SSD- 0.869 0.860
Imp. Faster R-CNN- 0.895 0.884
RetinaNet- 0.907 0.890
Experiment 1k1= k2 = k3 = 1 0.922 0.901
Experiment 2k1= 2, k2 = k3 = 1 0.917 0.902
Experiment 3k1= k3 = 1, k2 = 2 0.923 0.900
Experiment 4k1= k2 = 1, k3 = 2 0.918 0.901
Experiment 5k1= k2 = 2, k3 = 1 0.921 0.904
Experiment 6k1= k3 = 2, k2 = 1 0.914 0.902
Experiment 7k2= k3 = 2, k1 = 1 0.917 0.899
Experiments 8*k1= 0.0553, k2 = 0.3004, k3 = 0.6442 0.949 -
Experiments 9**k1= 0.0196, k2 = 0.1602, k3 = 0.8200 - 0.921
* Optimal Coefficients for 70% Training—30% Testing ** Optimal Coefficients for 80% Training—20% Testing.
Table 3. Success of diverse models on the NWPU-VHR10 dataset.
Table 3. Success of diverse models on the NWPU-VHR10 dataset.
ClassRICNN [2]CACNN [6]DFCCNN [7]FMSSD [8]COPD [10]RICAO [11]DODN [12]Improved Faster R-CNN [13]Our
Study
Plane0.88350.99910.90850.99700.62250.99700.93920.99971.0000
Ship0.77340.90550.90110.89900.69370.90800.92970.94580.9780
Storage Tank0.85270.90010.87680.90300.64520.90610.99250.36700.6982
Baseball Diamond0.88120.99650.98820.98200.82130.92910.96330.96910.9680
Tennis Court0.40830.90160.89500.86000.34130.90290.96120.93060.9310
Basketball Court0.58450.90910.90780.96800.35250.80130.70970.92950.9681
Ground Track Field0.86730.90910.90620.99600.84210.90811.00000.99890.9970
Harbor0.68600.88970.88720.75600.56310.80290.96880.93811.0000
Bridge0.61510.79620.90340.80100.16430.68530.81150.68710.9030
Vehicle0.71100.89000.87730.88200.44280.87140.80090.93280.9340
mAP0.72630.90970.90520.90400.54890.87120.90770.87100.9420
Table 4. Success of diverse models on the RSOD dataset.
Table 4. Success of diverse models on the RSOD dataset.
ClassSigNMS [5]A2RMNet [14]DConvNet [15]RFN [16]Our Study
Aircraft0.8060.9420.7180.7910.947
Oil Tank0.9060.9640.9030.9050.936
Overpass0.8740.8380.8951.0000.912
Playground0.9910.9970.9980.9971.000
mAP0.8940.9350.8790.9230.949

Share and Cite

MDPI and ACS Style

Körez, A.; Barışçı, N.; Çetin, A.; Ergün, U. Weighted Ensemble Object Detection with Optimized Coefficients for Remote Sensing Images. ISPRS Int. J. Geo-Inf. 2020, 9, 370. https://doi.org/10.3390/ijgi9060370

AMA Style

Körez A, Barışçı N, Çetin A, Ergün U. Weighted Ensemble Object Detection with Optimized Coefficients for Remote Sensing Images. ISPRS International Journal of Geo-Information. 2020; 9(6):370. https://doi.org/10.3390/ijgi9060370

Chicago/Turabian Style

Körez, Atakan, Necaattin Barışçı, Aydın Çetin, and Uçman Ergün. 2020. "Weighted Ensemble Object Detection with Optimized Coefficients for Remote Sensing Images" ISPRS International Journal of Geo-Information 9, no. 6: 370. https://doi.org/10.3390/ijgi9060370

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop