1. Introduction
Driven by the improved accessibility to remote sensing images, many deep convolutional neural network (CNN)-based object detection approaches have gained more attention recently. The successes of the deep detectors on remote sensing images facilitate a wide range of civil and military applications, such as urban planning [1], surveillance [2], crop monitoring [3], traffic planning [4], vehicle tracking [5], building detection [6,7], and military uses [8]. For the research purposes, several remote sensing object detection datasets such as DOTA [9], HRSC2016 [10], and NWPU VHR-10 [11] have been released and used in numerous works. Object detection with such public remote sensing datasets gives rise to following challenges: (1) since remote sensing images are obtained from bird’s-eye view, the restricted visual features of objects are captured; (2) objects appear with large scales variability in remote sensing images; and (3) the geospatial distributions of objects are highly imbalanced across multiple images as well as across different categories.
However, the public remote sensing object detection datasets are gathered from limited regions with some pre-defined categories, which restricts the flexible research usage in part. We construct our own datasets with 963 WorldView-3 images and 457 SkySat images covering harbor and airport peripheral areas, provided by DigitalGlobe [12] and Planet Labs [13] respectively. In our datasets, various types of aircraft and ship objects are labeled by experts in aerial image interpretation. Prior to directly exploiting the raw remote sensing images as the object detection datasets, it is worth noting the following conditions. First, the raw remote sensing images are very large in size (e.g., typically larger than 5000 × 5000 pixels). In public datasets, the raw aerial images were preprocessed into relatively small chips in general, so that it can be conveniently used for both training and testing. Secondly, the objects in our datasets are labeled in varying degrees depending on the amount of visual information and the resolutions of images, which leads to hierarchically partial annotations. Figure 1 shows some examples that one aircraft is labeled “aircraft (coarse1)|civilian-aircraft (coarse2)”, while the other aircraft are more elaborately annotated as “aircraft (coarse1)|civilian-aircraft (coarse2)|B-747 (fine)” or “aircraft (coarse1)|civilian-aircraft (coarse2)|A-380 (fine)”.
Due to the aforementioned first issue, given a huge remote sensing image we must explore the whole pixels during inference, suffering from high computational costs. In order to compensate the inefficient exploring process, we need an intensive region proposal strategy before performing the object detection. If ships are target objects, sea areas or shorelines would be of great importance in the ship detection rather than fully inland areas. Figure 2 provides two examples of actual remote sensing images taken from harbor-peripheral areas, where the ship objects are sparsely distributed in the aerial images. Inspired by this motivation, some researchers have been presented the region proposal methods for selectively analyzing regions on vast areas [14,15]. CRPN [14] generates the cropping regions by uniting boundaries of inferred object proposals, and CPNet [16] infers the candidate region proposals, trained with cluster groundtruth annotations generated by mean-shift algorithm [15]. We propose a novel region cropping strategy partially inspired by CRPN and CPNet. We also infer object proposals similarly to CRPN and after that, we perform the mean-shift clustering algorithm on the inferred object proposals. Unlike CPNet, we are not required to generate cluster groundtruth annotations. Also, we do not need to set any hyper-parameter because the mean-shift algorithm automatically finds the appropriate number of clusters fitted with data instances; meanwhile, CRPN sets a scale threshold for the generation of cropping regions.
On the other hand, handling the hierarchically partial-annotated objects mentioned as the second issue needs help basically from hierarchy-based detection methods. However, as majority of previous studies have been developed with the benchmark object detection datasets, where the objects are annotated with pre-defined fine labels arranged in a “flat” structure. Little research considers such datasets with hierarchically partial-annotated objects in object detection. Although it does not exactly match our dataset setting, there have been several works dealing with CNN-based hierarchical classification on natural images [17,18]. HD-CNN [17] learns classification components on coarse and subset of fine labels at the two levels and outputs a final probability distribution over fine labels. Ouyang et al. in [18] finetunes deep models, where a deep model for a group of classes is initialized with its parent deep model. Sharing the similar spirit as works [17,18], we devise a Hierarchical Sibling Classification Network (HSCN) by separately deploying classification modules on each set of sibling classes at all levels. We exploit the classification modules that are simple yet fully dense, instead of deep models as in [18]. Our approach leads to an end-to-end learning without the consecutively finetuning processes. Additionally, we minimize unnecessary competitions of a SoftMax function across all classes by replacing the SoftMax function with multiple SoftMax functions over sets of sibling classes when training HSCN, which differentiates our modeling from the approach of [17].
In this paper, we propose a hierarchical multi-label object detection framework for remote sensing images with the hierarchically partial-annotated datasets. This framework takes a raw remote sensing image of an arbitrary size as input, extracts interesting regions of the image, and then performs hierarchical multi-label object detection on the regions. The proposed framework consists of two parts. The first part is a clustering-guided cropping strategy, which quickly infers candidate object proposals and generates region crops tightening the clustered object proposals. Because we cannot speculate the exact position of objects before detection inference, the existing approach attempted to divide the whole aerial scene into sub-patch images and run inference over the all patch images. Our proposed clustering-guided cropping strategy tackles the inefficient detection process and resolves the problem through the shrinkage of exploration space via expedite objectness inference and clustering. Clustering on the roughly inferred objects leads to generating cluster crops which are valuable regions among whole scene areas to be intensively inferred. In the case that the time of decision making is a critical factor, our clustering-guided cropping strategy would play an important role in improving inference speed without sacrificing significant performance.
The second part is a novel hierarchical multi-label object detection pipeline called Decoupled Hierarchical Classification Refinement (DHCR), which involves two networks; one is a detection network with multiple classifiers and the other is a classification network called HSCN for hierarchical multi-labeling. The key common modelings for the two networks are locally deployment of classifiers for adapting to our hierarchically partial-annotated datasets. We are specifically inspired by the work of decoupled classification refinement (DCR) [19], which jointly accompanies a detection network and a classification network. The advantages of our proposed object detection pipeline called DHCR are: (1) DHCR can handle hierarchically partial-annotated dataset through the separation of classifiers, (2) DHCR can classify objects hierarchically through hierarchical classification results, and (3) DHCR can expect the improvement of classification accuracy at fine level as in DCR.
We carry out experiments on our own object detection datasets acquired from WorldView-3, SkySat satellites. For verifying the validity of the proposed framework, we qualitatively analyze by visualizing the experimental results, and quantitatively evaluate the performance with several rotational-box-based detector baselines. DHCR-based detections significantly boost the performance of the respective baseline models and we achieve state-of-the-art results on our own datasets. In summary, our contributions are four-fold:
- We are the first to tackle the hierarchical multi-label object detection problem under the hierarchically partial-annotated dataset settings.
- We propose a clustering-guided cropping strategy which infers object proposals via a deep detector and followed by clustering the object proposals by a clustering algorithm, which supports more efficient testing in subsequent object detection.
- We introduce a hierarchical multi-label object detection pipeline, DHCR, that are composed of a detection network with multiple classifiers and a classification network for hierarchical multi-labeling to train on arbitrarily hierarchical-annotated datasets and improve the performance of object detection.
- The extensive qualitative and quantitative experiments on our own datasets with WorldView-3 and SkySat satellite images suggest that the proposed object detection framework is a practical solution for cropping regions and detecting objects with hierarchical multi-labels.
The rest of this paper is organized as follows. Section 2 reviews related works on (1) object detection and applications of CNN in remote sensing images, (2) region proposal methods for aerial images, and (3) CNN-based hierarchical multi-label classification. The detailed design of our proposed object detection framework is described in Section 3. Section 4 reports comparative experimental results for the proposed object detection framework on our own datasets. Finally, Section 6 includes the final remarks and a conclusion.
2. Previous Work We first introduce object detection and applications of CNN in remote sensing images. Then we review two lines of works that are the most relevant to ours; region proposal methods for aerial images, and CNN-based hierarchical multi-label classification. 2.1. Object Detection and Applications of CNN in Remote Sensing Images
Due to rapid deep learning breakthroughs, CNN-based detection models have also drawn interest in the field of object detection on remote sensing images. In order to detect and count cars in unmanned aerial vehicle images, Ammour et al. [20] extracted features of regions through pre-trained CNNs and classify whether the regions are car via a linear support vector machine [21]. Another recent studies for vehicle detection are: Audebert et al. [22] proposed a deep-learning-based segment-before-detect method for segmentation followed by detection and classification, and Zhong et al. [5] combined two networks; one network generates a set of vehicle-like regions, and the other network takes a cascaded hierarchical feature learning approach. As military and civilian infrastructures, airports have been also studied for object detection [23,24]. For detecting ships with different scales and large appearance variation, the previous works [2,25] developed variants of Feature Pyramid Network [26], where Yang et al. built high-level semantic feature-maps for all scales [25], and Fu et al. [2] used deep reinforcement learning to get accurate ship angle information. As another application, Chen et al. [27] aimed at end-to-end airplane detection with a single deep CNN based on transfer learning. A few works have been proposed to detect fir trees for the forest health assessment in a quicker and cheaper way [3], and detect oil palm trees for predicting the yield of palm oil, monitoring the process of growth and resulting in the maximum productivity [28].
2.2. Region Proposal Methods for Aerial Images
The aerial images generally cover vain areas and such characteristic has been motivated many researchers to develop region proposal methods for focusing on regions selectively. The purposes of previous works on the region proposal methods can be viewed from two aspects. First, object detection needs to be conducted on certain areas excluding the massive backgrounds with very little chance of target objects being present [14,29]. Secondly, because objects taken in the aerial images are very small in pixels and distributed sparsely and non-uniformly, e.g., pedestrians or vehicles, object detectors can show improvement by re-executing object detection especially in zoom-out images whose target objects are densely crowded [16,30].
For the former purpose, the work in [14] proposed a Cropping Region Proposal Network (CRPN), which exploits a weak semantic RPN quickly searching for candidate object proposals and unites neighboring objects based on Intersection-over-Union (IoU). The other work sharing the same purpose [29] predicted whether objects exist or not for a patch image generated by a sliding-window fashion. Given in a patch image, Pang et al. [29] extracted a feature vector on the patch image via a lightweight residual network called Tiny-Net and a classifier takes the feature vector for the binary objectness prediction. For the latter purpose, a Density Map guided-detection Network called DMNet was presented in [30], which estimates a density map for a given input aerial image and crops connected regions based on the estimated density map. To be specific, DMNet obtains a density mask by applying a density threshold to the estimated density map and uses connected component algorithm to form the cropping connected regions. Another work [16] proposed a Cluster Proposal sub-Network called CPNet which predicts cropping regions themselves. There is a clear distinction between CRPN and CPNet, where CRPN infers object proposals followed by unity of the boundaries of neighboring objects while CPNet infers region proposals. CPNet uses a clustering algorithm called mean-shift [15] for construction of the cluster bounding box annotations as its training dataset.
2.3. CNN-based Hierarchical Multi-Label Classification
There have been a few approaches to address hierarchical multi-label classification problem on natural images rather than aerial images [17,18,31,32,33]. This line of work was pioneered by HD-CNN [17] which embeds deep CNNs for learning the two-level organization of coarse and fine categories. HD-CNN deployed independent classification components for a set of coarse and each subset of fine categories. Training the classifier on fine categories depends on the low-level features which are class-agnostic such as corners and edges, and is conditionally executed to coarse category consistency and coarse predictions. Ouyang et al. [18] performed cascaded hierarchical feature learning for object detection. They constituted multiple groups of object classes into hierarchical clusters and used a deep model on each cluster for the hierarchical feature learning. The deep models are finetuned using their parent models as initial points. YOLO9000 [31] resolved hierarchical classification problem simply by multiplying a series of conditional probabilities through partitioned SoftMax operations, where YOLO9000 is an improved version of YOLO [34]. YOLOs [34,35] are representative one-stage object detectors that have been successfully used to address video summarization problem [36,37]. Branch-CNN (B-CNN) [32] took a slightly different approach, with sequential classification modules on the set of classes from coarse to fine levels. The feature vector for a given image extracted from a parent classification module is propagated to the next child classification module. More recently, Parag and Wang [33] proposed a multi-layer dense connectivity for a CNN to classify both the category and the ordered chain of ancestor classes. They devised the hierarchical arrangement of dense connections, where the hidden nodes for a child class are outputted by corresponding dense layers and re-weighted by the predicted probability for the child class.
3. Proposed Methodology
Even experts does not always guarantee the fully detailed annotations, as shown in Figure 1 because the aerial images are usually obtained from a top-down view, capturing the restricted visual appearances of objects. For example, there is an aircraft labeled as “aircraft (coarse1)|civilian-aircraft (coarse2)”, while another aircraft is identified up to “aircraft (coarse1)|civilian-aircraft (coarse2)|B-747 or A-380 (fine)”. We propose a hierarchical multi-label object detection framework in remote sensing images, which uses such dataset with the mixed labeling levels. The proposed framework is mainly made up of two parts: (1) clustering-guided cropping strategy, which takes a vast aerial image as an input and specifies the target regions to be subsequently processed by the object detection and (2) hierarchical multi-label object detection over the target regions by decoupled hierarchical classification refinement (DHCR). In this section, we describe details of our proposed framework including overview, clustering-guided cropping strategy, and DHCR.
3.1. Overview of the Proposed Hierarchical Multi-Label Object Detection Framework In contrast to natural imagery collected on the ground, the generic remote sensing imagery obtained by electro-optical sensors covers massive areas. Therefore, to perform inference while scanning remote sensing images, we should crop the large-scaled remote sensing images into the set of patch images small enough to enter the detection network. This cropping process brings the larger increase of time and memory cost with the greater number of patch images including only meaningless information, e.g., background images not containing any objects. We need to carefully filter out the unnecessary regions without losing target objects.
The overview of our proposed framework is illustrated in Figure 3. Inspired by the above-mentioned motivation, we first extract the regions required for object detection, which is most likely to include target objects. We call this process clustering-guided region proposal. Next, we create a series of patch images from the cluster crops in a sliding-window fashion. We then feed the patch images into our proposed hierarchical multi-label object detection pipeline. For each input patch image, we perform the inference of object detection, where the inference results are localized objects with corresponding hierarchical multi-labels.
3.2. Clustering-Guided Cropping Strategy
Since we cannot conjecture the approximate position of objects before inference, the existing approach tried to split the whole region of an aerial image into the set of patch images as shown in the left side of Figure 4, which we call uniform cropping throughout this paper. For example, in the case of ship detection, analyzing inland areas becomes less important compared to sea sides, island areas, or oceans with floating objects. The essence of our proposed framework is to first extract the interesting regions and then run the object detection on the extracted regions. We propose a simple yet efficient cropping strategy called clustering-guided cropping for the efficient detection, as illustrated in the right side of Figure 4.
We now elaborate the design of clustering-guided cropping strategy; the overview is depicted in Figure 5. The key idea of our proposed strategy is to infer object proposals by quickly scanning input aerial image and determine the boundary of cluster crops by gathering and clustering the inferred object proposals. For generating the candidate object proposals where objects are likely to be present, we use Region Proposal Network (RPN) which is a well-known sub-network of typical two-stage detectors.
To create dataset for training and testing RPN, we split the input aerial image evenly with sizeM×M′normally larger than the input patch sizeN×N′of the object detector. A larger value of M leads to smaller number of split images and faster inference of RPN; however, the excessive size M would cause a deterioration of the objectness accuracy. The reason is thatM×M′patch should be resized to a pre-defined input size for entering into RPN, due to GPU memory constraints. As the value of M increases, the spatial resolution of the resized image is more decreased, and objects become smaller. We should set M in consideration of the speed-accuracy trade-off. With the usage of RPN that is robust to detection of small objects such as feature pyramid network (FPN), we could deal with the small object detection despite the relatively large M value. We adopt RPN for faster inference due to the lightweight structure consisting fully convolution layers, but are not limited to the usage of RPN for object proposals.
After inferring the objectness via RPN, we merge all detected objects and run min-shift clustering algorithm [15], following the clustering process of groundtruth object boxes in [16]. The min-shift clustering algorithm is a nonparametric centroid-based clustering algorithm taking the density of data instances into account, without any prior knowledge of the number of clusters. We cluster the center points of the detected boxes by the min-shift clustering algorithm, and then get the set of minimum and maximum coordinates of boxes belonging to the same cluster. Finally, we determine the boundary of each cluster crop as a pair of the minimum and maximum coordinates—the top-left corner and bottom-right corner. We split each cluster crop into the set ofN×N′patch images in a sliding-window manner. Later, the patch images are fed into the proposed object detection pipeline. The detailed implementation for clustering-guided cropping strategy is summarized in Algorithm 1:
Algorithm 1: Clustering-Guided Cropping Strategy |
Input: a remote sensing image I, a score thresholdτOutput: coordinates of cluster cropsR
/* 1. Inference of Candidate Object Proposals via RPN */
|
3.3. Decoupled Hierarchical Classification Refinement
3.3.1. Pipeline Design
We propose a novel object detection pipeline called Decoupled Hierarchical Classification Refinement (DHCR), which explicitly adapts to hierarchically partial-annotated dataset depicted in Figure 1. We build this pipeline through combining two networks in a sequential manner as shown in Figure 6: (1) an Object Detection network with Multiple Classifiers (ODMC); (2) a Hierarchical Sibling Classification Network (HSCN). Figure 6 shows our pipeline design. The separated deployment of a detection network and a classification network is inspired by the success of decoupled classification refinement (DCR) [19]. Beyond the detection of objects labeled in a “flat” structure, we extend DCR to the hierarchical multi-label object detection.
When an input patch image is fed into ODMC, objects are detected in the form of bounding boxes with corresponding confidence scores over categories. Then, the objects are warped to a fixed size (e.g., 224 × 224) for forwarding them to HSCN, where HSCN substitutes the traditional single classification branch with multiple classification branches for outputting hierarchical multi-label classification results. The final score of HSCN for an input object combines confidence scores from all outputs of deployed classifiers. Finally, we merge the predicted probability distributions from ODMC and HSCN by element-wise multiplication.
As shown in Figure 6, we have following core modules in our proposed pipeline:
- RoIProposals from ODMC: As a typical module in general two-stage detectors, this module generates RoIs likely to include target objects on an input patch image.
- Regressor from ODMC: For RoIs from RoIProposals from ODMC module, this module corrects localization boxes for each RoI. After this bounding box refinement, we create input RoI images of HSCN in an upright position.
- Classifiers from ODMC: For RoIs from RoIProposals from ODMC module, this module estimates over complete and incomplete categories at fine level based on the results from multiple classifiers. According to the predicted results of this module, false negative boxes are determined and appended to the input RoI images of HSCN.
- Classifiers from HSCN: This module not only refines the fine-level classification result of Classifiers from ODMC module, but also predicts probability distributions over categories at each coarse level.
The common approach of ODMC and HSCN lies in replacing the single classifier with multiple parallelized classifiers. In the final layer of a typical classification module, the representative operation, e.g., SoftMax, is usually used for computing a probability distribution over all classes. In other words, the probabilities across all classes compete against each other and optimization might be biased depending on the category frequency [39]. In the case of HSCN, given the hierarchically annotated dataset where there are hierarchical relationships among classes, we partition the global competition over all classes into the local competitions between sibling classes. We can learn the classifier weights specialized in each set of sibling classes. ODMC takes more simplified approach; we define two kinds of class sets as described in the top side of Figure 6 and learn the two class sets separately.
We divide the classifiers in ODMC into two main types; (1) the SoftMax classifier over complete-categories, and (2) the sigmoid/SoftMax classifier(s) over subsets of incomplete-categories. The former classifier is a generic classifier in existing object detection networks, while the latter classifier is additionally modeled with sigmoid or SoftMax operations. In the case of the sigmoid operation, we can exploit multi-labels with not only a fine label but also its parent coarse label(s). Because of the sigmoid properties, the coarse label(s) even in complete-categories can be used without the competition burden unlike SoftMax. We devise the classification architecture of ODMC taking the hierarchically partial-annotated objects into account to enable the reasonable learning: with a single SoftMax classifier, it is not proper for simultaneously classifying over complete-categories and incomplete-categories, because complete-categories and incomplete-categories are hierarchically related, not completely exclusive.
Analogously, HSCN builds multiple classifiers instead of a single classifier. We involve the multiple classifiers for each set of sibling classes at all levels. The multiple classifiers share the same feature vector extracted from the backbone CNN network and each classifier has the same classification module;fc-ReLU-fc layers followed by SoftMax, as shown in Figure 6. The intuition behind this modeling approach is that we formulate a hierarchical multi-label classification problem as multiple flat classification problems defined over sets of sibling classes. As for the background class, we include it only in the most top classifier, i.e., the classifier over coarse1 labels, since duplicate trainings in other classifiers are unnecessary. The final output probability over a fine class is estimated by multiplying probabilities for its parent classes recursively, which more details will be further described in Section 3.3.3.
3.3.2. Training
The proposed pipeline requires a two-stage training process, where the first stage is learning ODMC, and then using the output of the first stage, HSCN is trained in the second stage sequentially. All the predicted bounding boxes including false negatives are warped to a fixed size (e.g., 224 × 224) which are fed into HSCN. If the boxes are not horizontal but rotational, then the boxes should be rotated in an upright position before warping. We re-train the boxes with the strong classifier HSCN and then support the classification prediction results by ODMC. ODMC jointly optimizes the regression and the classification, and hence may lead to sub-optimal for the respective problems [19]. We aim to make HSCN train on the misleading prediction domain generated from ODMC.
ODMC attaches additional classifier(s) to an extracted RoI-wise feature vector, which parallels the original SoftMax classifier. As mentioned in Figure 6, we divide categories into the aforementioned two types of class sets: complete-categories and incomplete-categories. ODMC separately optimizes for complete-categories and incomplete-categories by the original SoftMax classifier and the other additional classifier(s), respectively. For the newly added classifier(s), we introduce two different methods: a SoftMax classifier trained with a single label or a sigmoid classifier trained with multiple labels. We should consider the characteristics of dataset for the appropriate choice of classifier methods and the composition of the classifiers.
We use a multi-task loss to minimize the objective function for the detector of ODMC:LODMC=Lreg+Lcls, whereLregandLcls are loss functions for the classification and the regression layer, respectively. Here, our classification loss function for a labeled RoI is differently defined according to the type of the classifier(s) as Equations (1),2:
With SoftMax classifier(s),
Lcls(pα,yα,pβ,yβ)=LSCE(pα,yα)+∑iλi[yiβ∈Ciβ]LSCE(piβ,yiβ)
with sigmoid classifier(s),
Lcls(pα,yα,pβ,yβ,yparβ)=LSCE(pα,yα)+∑iλi[yiβ∈Ciβ]LBCE(piβ,{yiβ,ypar(i)β})
whereSCEandBCEspecify the SoftMax/sigmoid cross-entropy loss.pα,piβdenote the probability distribution over complete-categories andithsubset of incomplete-categories, respectively.yα,yiβdenote the groundtruth class belonging to complete-categories andithsubset of incomplete-categories, respectively. If any groundtruth category of the labeled RoI does not belong toithsubset of incomplete-categories, i.e.,yiβ∉Ciβ, the Iverson bracket indicator function[yiβ∈Ciβ]evaluates to 0 and 1 otherwise. For each sigmoid classifier trainable with multi-labels, we can arbitrarily use parent classes ofyiβdenoted byypar(i)β, required for training the classifier.λiis the balancing parameter for classification onithsubset of incomplete-categories.
HSCN addresses the hierarchical multi-label classification problem by locally placing flat classification modules over the sets of sibling classes. As for the background class, we included in the most top classifier for competing only against the coarse1 labels. All localized classification modules share the same image-level feature vector as input data. The presented architecture pursues a flexible training for sub-classes without the being constrained to hierarchical dependency. We design the loss function of HSCN for each warped RoI to be a weighted summation of all SoftMax cross-entropy losses as in Equation (3).
L({ps,ys}s=1S)=∑s⊂Sλs ts LSCE(ps,ys)
where S is the union of sets of sibling classes at all levels and s is a set of sibling classes as a subset of S.tsdenotes whether a groudtruth label y belongs to s. If a groundtruth label y does not belong to s, then we cannot defineysand do not reflect corresponding loss,LSCE(ps,ys)by settingts=0, otherwise the valuetsis 1. For each sibling class group s, the SoftMax cross-entropy loss is optimized with the groundtruth labelyson the probability distributionps.λsis the balancing weight for the classifier over s.
3.3.3. Inference
Inference is also processed in a sequential manner. For each input image, ODMC first outputs localized objects with predicted labels by SoftMax-SoftMax classifiers or SoftMax-sigmoid classifiers. The all inferred bounding boxes of objects are warped to a fixed size. Afterwards, HSCN takes the warped bounding boxes as inputs and estimates the probability distributions over sibling classes at all levels. HSCN computes the probability for a fine label by recursively multiplication of all ancestor probabilities including itself. We obtain an unnormalized final probability distribution over all fine labels through element-wise product of the two probability distributions from ODMC and HSCN. The above all computations for probability distributions over fine labels for each labeled RoI can be expressed by the following Equations (4)–(6).
pj,LODMC=[yj,L∈Cα]pj,Lα+[yj,L∉Cα]ppar(j),l*α pj,Lβ· ∏l=l*+1L−1ppar(j),lβ·
pj,LHSCN=pj,L∏l=1L−1ppar(j),l
pj,LDHCR∝pj,LODMC·pj,LHSCN
whereyj,lis the groundtruth class at level l.pj,L(·)is the final predicted probability given a model foryj,L. The Iverson bracket indicator functions[yj,L∈Cα]and[yj,L∉Cα]evaluate to 1 if theyj,Lis in/not in complete-categories, respectively.pj,lβ· denotes the probability for class j at level l from corresponding classifier,β·. Here,par(j),ldenotes the parent class at level of class j.
4. Experiments
To validate the contributions of the proposed framework, we conduct extensive experiments on 1) clustering-guided cropping strategy in Section 4.2 and Section 2) decoupled hierarchical classification refinement in Section 4.3. In this section, we start by explaining our datasets used for the experiments in Section 4.1, and then present implementation details and various qualitative and quantitative experimental results.
4.1. Dataset Description
We constructed two datasets with 963 WorldView-3 [12] and 457 SkySat [13] pansharpenend remote sensing images, mainly covering harbor- and airport- peripheral areas. Each remote sensing image is labeled by experts in aerial image interpretation with multiple categories, and the categories are hierarchically organized. The category hierarchy and the number of objects for each category are given in Figure 7. For WorldView-3 satellite, the images have a resolution of 0.3 m, and is of the size in the range from about 200 × 7000 to 16,000 × 16,000 pixels (sorted by their areas). For SkySat satellite, the images have a resolution of 1 m, and is of the size in the range from about 22,000 × 8000 to 23,000 × 31,000 pixels (sorted by their areas). We randomly split satellite images into 60% for training, 20% for validation, and 20% for testing on each dataset, and then create patches from all the satellite images to avoid duplicate object information being included in both training and testing patch images. In our datasets, complete-categories are CA, MA, CO, OT, MV, AC, SU, LW, and SW. The short names for the categories are defined as: CO=COntainer, OT=Oil Tanker, MV=Maritime Vessels, AC=Aircraft Carrier, SU=SUbmarine, LW=Large Warship, and SW=Small Warship. In our datasets, only some objects are annotated up to specific aircraft such as civilian aircraft (e.g., B-707, B-717, B-727, etc.) and military aircraft (e.g., A-10, A-12, A-18, etc.).
4.2. Experiments on Clustering-Guided Cropping Strategy We perform qualitative experiments and a series of sensitivity analyses on the WorldView-3 dataset. For the qualitative experiments, we visualized some example results of cluster crops generated on the WorldView-3 test images. The sensitivity analyses are conducted to find appropriate settings for RPN and construction of patch images from the generated cluster crops.
4.2.1. Implementation Details
We use RPN of Feature Pyramid Network (FPN) (Re-implementation version publicly from https://github.com/DetectionTeamUCAS/FPN_Tensorflow) [26] for inferring candidate object proposals, which is robust to variance in scale. All input patch images are entered into RPN after resizing to1024×1024 . We follow the original source code on hyper-parameters unless specified. Specifically, MomentumOptimizer [40] with momentum of 0.9 is used for optimization. We train a total of 150 k iterations, with a learning rate of 0.001 for the first 60 k iterations, 0.0001 for the next 20 k iterations, and 0.00001 for the remaining 70 k iterations. We use a 5-level pyramid levels from P2 to P6, and at each pyramid level, anchors with one scale {1.0} and three anchor ratios {0.5, 1.0, 2.0} are used. All evaluations were carried out on two TITAN Xp GPUs with 12 G memory; one GPU for ODMC, and another GPU for HSCN.
4.2.2. Qualitative Results
We generate cluster crops on the WorldView-3 test datasets and the results on harbor and airport peripheral areas are displayed in Figure 8 and Figure 9, respectively. The top side of each figure presents the image with the groundtruth objects represented by yellow bounding boxes. The bottom side of each figure provides the image with the detected objects and the generated cluster crops, plotted in different colors according to the clusters. We use the raw patch size of 3000 before down-scaling to a fixed input size of 1024 when training and testing RPN. For running mean-shift clustering, we only take object proposals whose objectness scores are higher than 0.985.
In the example of Figure 8, three cluster crops are generated. The regions suggested by the cluster crops are mainly spread to the sea areas including the sea shores. This experimental result fits our intuition; the target objects as ships are likely to be in certain areas such as sea sides, island areas, and oceans rather than inland areas. As shown in Figure 8a although there does not groundtruth objects in the sea shore, our RPN infers candidate object proposals due to small-scale traces which look like ships or very small ships that have not yet been unlabeled. However, as we are in the stage of the region proposal for subsequent object detection, the suggested sea shore regions are still plausible as the candidate regions to be explored. Figure 9 presents the other experimental result, which also generates three cluster crops. Many aircraft are detected as candidate object proposals especially around aerial routes.
Overall, we see that the generated cluster crops in Figure 8b and Figure 9b well capture the regions, in which the groundtruth objects are mostly situated. Massive backgrounds with very little chance of target objects being present take up many parts in the other regions that are not covered in the generated crops. Our clustering-guided cropping strategy helps saving the cost of exploring the less important regions.
4.2.3. Sensitivity Analysis
We perform the sensitivity analyses varying (1) the raw patch sizes of {3000, 5000, 7000, 9000} before down-scaling to a fixed input size of 1024 when training and testing RPN, (2) the confidence thresholds of {0.980, 0.985, 0.990, 0.995} for clustering only objects that are detected with a confidence score higher than the confidence threshold, and (3) the padding sizes of {300, 600, 900} from outside crop border line. We use four metrics and details are described as follows:
- Number of patches to be inferred: For uniform cropping, we count the number of patches created from the whole images, while for clustering-guided cropping, we count the number of patches created from the cluster crops.
- Number of groundtruth objects: the number of groundtruth objects included in the output cluster crops. In the case of uniform cropping, there is no loss to the groundtruth objects unless truncated.
- Total inference time (s/img): the average running time per image from the beginning of the object detection framework until the end.
- Average of mean Average Precision (Avg. of mAP) (%): detection performance over incomplete-categories. mAP is used as a standard metric to evaluate the performance of object detection and computed as the average value of AP over all categories. Here, AP computes the average value of precision over the interval from recall = 0 to recall = 1. The precision measures the fraction of detections that are true positives, while the recall measures the fraction of positives that are correctly identified. Hence, the higher the mAP, the better the performance.
Through the metrics, we explore the trade-off between time and memory cost (measured with the first and third metrics), and accuracy (measured with the second and fourth metrics) compared with uniform cropping. The uniform cropping illustrated in Figure 4 splits the whole region of an aerial image into the set of patch images, while clustering-guided cropping splits the regions selectively. Therefore, the clustering-guided cropping cannot completely surpass the uniform cropping statistically only in terms of accuracy. We ultimately aim at efficient object detection guided by our conjecture of regions likely to include target objects, without loss of groundtruth objects as much as possible.
The sensitivity analyses are carried out to find out the adequate hyper-parameters such as the raw patch size before down-scaling, the confidence threshold for clustering, and the padding size from outside crop border line. As the larger the raw patch size, the lower the image resolution and the smaller the object sizes. The results in Table 1a demonstrate the impact of the raw patch size, where experimental cells with 3000 × 3000 show the best mAP on average and mAP gradually degrades as the raw patch size increases. It is noteworthy that the total inference time is cut in about half on 9000 × 9000 compared to 3000 × 3000 while the average performance only drops by 2.69%. From Table 1b, we can see that the confidence threshold is another influential factor. The higher we set the confidence threshold, the fewer objects are taken, which leads to decreasing the total runtime but causing performance drop because the likelihood of losing groundtruth objects is getting higher. In the case of the third sensitivity analysis in Table 1c, we differentiate the padding size by 300, 600, and 900; since the region proposals tighten the boundaries of the clustered object proposals, we reserve margins through the paddings so as to relax the dependency on inferred object boundaries. The adequate padding size obviously may help to improve the performance; however at least in our experiments, it yields gaps that are not that large.
We investigate top-N experimental cells by sorting in descending order of their detection performance, mAP, and plot the speed/accuracy trade-off curve in Figure 10. From Figure 10, we see that the total inference time steadily consumes, mAP values increase and are approaching mAP of uniform cropping. Although some experimental cells show similar performance as the uniform cropping, there are comparative benefits in terms of the total inference time. For some experimental cells surpassing mAP of the uniform cropping, we simply attribute the small gaps with the uniform cropping to difference in input data distribution.
4.3. Experiments on Decoupled Hierarchical Classification Refinement Considering a wide range of object scales of WorldView-3 and SkySat satellite images, we apply a multi-scale testing at the inference stage, and use testing scales of {768, 1024, 2048} with corresponding strides of {192, 256, 512}. The raw patch images of different sizes are then resized to have fixed sizes, 768 × 768, for entering the detection networks. The SoftMax operations are used in all the classifiers.
4.3.1. Baselines
Due to the characteristics of the aerial images, rotational bounding boxes are more suitable to encompass the objects and distinguish crowded objects from each other rather than horizontal bounding boxes [41,42]. We choose rotational object detection networks as baselines; (1) Rotation Region Proposal Networks (RRPN) [43] which learns with inclined rectangle proposals, (2) Rotational Region CNN (R2CNN) [44] which estimates both the axis-aligned box and the inclined minimum area box using different pooled features, (3) Detector for Small, Cluttered and Rotated objects (SCRDet) [45] which introduces a sampling fusion network, and pixel/channel attention networks, (4) a simple one-stage dense detector called RetinaNet [46] which optimizes focal loss, and 5) Rotated bounding BOX-based CNN (RBox-CNN) [47] which is based on rotation anchors and implements stable regression. All baseline detectors typically have a single SoftMax classifier at the end of each detection network. We learn the SoftMax classifier only over complete-categories. The SoftMax classifier is not proper for simultaneously classifying over complete-categories and incomplete-categories, because complete-categories and incomplete-categories are hierarchically dependent, not completely exclusive.
4.3.2. Implementation Details
We use the publicly released implementations for baseline models of RRPN (https://github.com/DetectionTeamUCAS/RRPN_Faster-RCNN_Tensorflow), R2CNN (https://github.com/DetectionTeamUCAS/R2CNN_Faster-RCNN_Tensorflow), SCRDet (https://github.com/DetectionTeamUCAS/R2CNN-Plus-Plus_Tensorflow), RetinaNet (https://github.com/DetectionTeamUCAS/RetinaNet_Tensorflow_Rotation), and RBox-CNN (https://github.com/SIAnalytics/simplified_rbox_cnn), where these models are implemented based on TensorFlow [48] framework. For all baseline models except for RetinaNet, we unify the backbone networks into ResNet-101 [49], feed input images sized of768×768 , and adopt the pre-trained weights on DOTA [9] that are publicly available (https://github.com/DetectionTeamUCAS/Models). When using RetinaNet, we adopt ResNet-50 [49] as backbone network, take input images sized of592×592 , and finetune over the pre-trained models (https://github.com/DetectionTeamUCAS/RetinaNet_Tensorflow_Rotation) on ImageNet. We follow the original source code on hyper-parameters unless specified.
For implementation of ODMCs, we use each baseline as a main body of ODMC and classification modules of the baselines are tailored to ODMC. We then build corresponding DHCR for each ODMC and each HSCN is trained with all the predicted bounding boxes from ODMC, including its false negatives. All bounding boxes are warped to a fixed size as224×224 and then fed into HSCN. For the experiments, we constitute the classification branches of ODMCs as (1) a SoftMax classifier over complete-categories, and (2) a SoftMax classifier over incomplete-categories. We implement HSCN based on the TensorFlow-Slim classification network (https://github.com/tensorflow/models/tree/master/research/slim). As with ODMCs, we replace the single SoftMax classifier in the downloaded source code with multiple SoftMax classifiers over the sets of sibling categories at all levels. All HSCN models are learned by stochastic gradient descent (SGD) optimizer and we train a total of 600 k iterations with a learning rate of 0.0005.
4.3.3. Quantitative Results
Table 2 and Table 3 demonstrate the effectiveness of DHCR over different backbones and frameworks. We present the experiments on WorldView-3 and SkySat datasets. Without the help of DHCR, the existing baseline models with a single classifier limit to classifying in a flat N-way over complete-categories such as CA, MA, CO, OT, MV, AC, SU, LW, and SW in our dataset. The other classes at the fine level are partially annotated to some objects, and therefore the objects are learned with their corresponding coarse2 labels. In order to compute the AP values for classes not belonging to complete-categories like S and A at coarse1 label, and CS and MS at coarse2 label, we intentionally summed the children probabilities for a parent class and assume the summed value as the probability of the parent class, e.g.,p(CS)=p(CO)+p(OT)+p(MV).
As can be seen in Table 2 and Table 3, DHCR-based detections significantly improve the performance of the respective baseline models. In the proposed pipeline named DHCR, the embedded detector is able to classify into multiple labels across different sets of labels and the accompanying classifier named HSCN plays an important role in improving the performance of detection. After applying DHCR to each baseline detection model, the level-wise performance (mAP at L1/L2/L3) are all further improved on WorldView-3 dataset: from 63.58/54.43/45.05 to 66.10/55.46/53.57 for RRPN, from 64.17/56.62/47.99 to 73.01/60.58/58.12 for R2CNN, from 64.74/58.27/48.37 to 74.75/61.76/58.25 for SCRDet, from 69.53/61.92/52.33 to 81.15/69.27/63.30 for RBox-CNN, and from 62.50/53.46/44.72 to 65.10/54.41/52.15 for RetinaNet. Experimental results on SkySat dataset also show similar patterns of the performance improvement. The full model with RBox-CNN and DHCR achieves the best performance. Most importantly, thanks to the flexible and general design of the proposed pipeline called DHCR, it is adaptable to any dataset with annotations with an arbitrary hierarchical structure and can be expected to enhance the performance by usage of the hierarchy information.
4.3.4. Qualitative Results
In Figure 11 and Figure 12, we visualize detection results on the WorldView-3 and SkySat test images, respectively. The detection results on both datasets commonly show substantial improvement in overall confidence scores of the true positives and detects objects with hierarchical multi-labels. Although only a few objects are labeled up to specific aircraft throughout our datasets, we could prevent a biased learning to major classes due to the key design factor in HSCN that locally deploys the classification modules. As shown in the right columns of Figure 11 and Figure 12, our pipeline presents the hierarchical multi-label object detection results, which provides much semantic and interpretable information about the objects.
5. Discussion
When we create input patch images of RPN, we must crop each huge scene image into lots of patches sized M. It is recommended that the hyper-parameter, M, be set to the value over the interval from the actual input size of RPN and to the overall scene size. A lower value of the actual input size of RPN leads to the slower speed and less efficient inference, and a higher value of the overall scene size results in including unnecessary empty areas and degrading the detection performance. In our case, the actual input size of RPN is 1,024 and the overall scene size is about 10,000. Therefore, we test M with diverse values from 3,000 to 9,000. Table 4 shows the effect of the raw input size, M, on the performance of RPN and total computational times for the training and inference process of DHCR. The hyper-parameter setting of M to a large value reduces the number of training and testing patch images generated from each scene image, which obviously brings about decreasing the computational times and declining the detection performance.
Together with Table 4, we also evaluate the average running time per patch image sized 768 × 768 for time complexity analysis among baseline models and ours in Table 5. As expected, the representative model of the one-stage detection model, RetinaNet [46], is the fastest, followed by RRPN [43], R2CNN [44], SCRDet [45], and RBox-CNNN [47]. Compared to baseline models, our model is measured as the slowest one because the backbone of ODMC is set to RBox-CNN showing the best detection performance in our experiments and we additionally perform the hierarchical classification refinement through the inference of HSCN.
Figure 13 shows the comparison of confusion matrices before (left) and after (right) the hierarchical classification refinement. In other words, the left result is achieved after the inference of only ODMC and the right result is obtained from DHCR which fuses the results of ODMC and HSCN. As shown in the right side of Figure 13, the detection scores for the overall categories are improved after reflecting the results of HSCN, which visually provides the quantitative contribution of HSCN.
6. Conclusions We have presented a flexible and general framework for hierarchical multi-label object detection for remote sensing images. The framework extracts crucial regions to be inferred in the subsequent object detection process on input aerial images via our clustering-guided cropping strategy, and then performs hierarchical multi-label object detection on the extracted regions by our detection pipeline called decoupled hierarchical classification refinement (DHCR). DHCR fuses the results of two networks: (1) an Object Detection network with Multiple Classifiers (ODMC), and (2) a Hierarchical Sibling Classification Network (HSCN) for supporting hierarchical multi-label classification. Due to the flexible and general design of the proposed pipeline, it is adaptable to any dataset with annotations with an arbitrary hierarchical structure. Our method shows significant improvements over several rotational detection baselines on our own two datasets constructed from WorldView-3 and SkySat satellite images. Hence, it demonstrates a practical solution for cropping regions and detecting objects with arbitrary hierarchically partial-annotated datasets. The current flexible and general design, however, requires users to adjust the model structure such as the specific determination of classification modules depending on the labeled classes of dataset, which needs to be further improved for convenient usage. As a future work, for much more realistic scenarios, we plan to further improve our proposed framework to address domain adaptation across image datasets from different satellites.
Raw Input Size (M) | # Patches | # Groundtruth | Total Inference | Avg. of mAP (%) |
---|---|---|---|---|
before Down-Scaling | to Be Inferred | Objects | Time (s/img) | |
Uniform Cropping | 47,510 | 24,371 | 64.07 | 61.40 |
3000 × 3000 | 27,803 | 22,634 | 45.46 | 60.10 |
5000 × 5000 | 24,477 | 21,511 | 40.42 | 59.28 |
7000 × 7000 | 23,331 | 21,165 | 38.76 | 59.03 |
9000 × 9000 | 18,346 | 19,584 | 31.33 | 57.41 |
a Sensitivity analysis varying the raw patch sizes before down-scaling to | ||||
a fixed input size of 1024 when training and testing RPN. | ||||
Confidence | # Patches | # Groundtruth | Total Inference | Avg. of mAP (%) |
Threshold of RPN | to be Inferred | Objects | Time (s/img) | |
Uniform Cropping | 47,510 | 24,371 | 64.07 | 61.40 |
0.980 | 25,957 | 21,928 | 42.38 | 59.49 |
0.985 | 24,295 | 21,672 | 40.03 | 59.21 |
0.990 | 22,399 | 20,616 | 37.48 | 58.50 |
0.995 | 19,440 | 20,294 | 33.40 | 58.08 |
b Sensitivity analysis varying the confidence thresholds for clustering only objects | ||||
that are detected with a confidence score higher than the confidence threshold. | ||||
Padding Size | # Patches | # Groundtruth | Total Inference | Avg. of mAP (%) |
of Cluster Crops | to be Inferred | Objects | Time (s/img) | |
Uniform Cropping | 47,510 | 24,371 | 64.07 | 61.40 |
300 | 19,075 | 21,127 | 33.05 | 58.53 |
600 | 22,760 | 21,127 | 37.89 | 58.56 |
900 | 27,231 | 21,127 | 44.04 | 59.36 |
c Sensitivity analysis varying the padding sizes from outside crop border line. |
Coarse1 Label (L1) | Coarse2 Label (L2) | Fine Label (L3) | ||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Base Detector | +DHCR | S | A | mAP | CS | MS | CA | MA | mAP | CO | OT | MV | AC | SU | LW | SW | CA1 | CA2 | MA1 | MA2 | mAP† | mAP |
RRPN [43] | ✗ | 52.23 | 74.92 | 63.58 | 62.72 | 41.12 | 70.13 | 43.74 | 54.43 | 65.38 | 53.56 | 52.81 | 32.94 | 45.68 | 35.01 | 29.94 | - | - | - | - | 45.05 | - |
R2CNN [44] | ✗ | 53.43 | 74.91 | 64.17 | 69.36 | 42.13 | 81.93 | 33.06 | 56.62 | 70.58 | 56.94 | 59.04 | 33.17 | 46.26 | 39.13 | 30.81 | - | - | - | - | 47.99 | - |
SCRDet [45] | ✗ | 54.77 | 74.71 | 64.74 | 69.91 | 42.95 | 84.54 | 35.67 | 58.27 | 71.26 | 55.62 | 60.99 | 34.99 | 47.22 | 40.41 | 28.13 | - | - | - | - | 48.37 | - |
RBox-CNN [47] | ✗ | 64.57 | 74.48 | 69.53 | 72.12 | 44.24 | 80.47 | 50.83 | 61.92 | 70.35 | 75.37 | 61.37 | 34.04 | 50.88 | 40.05 | 34.24 | - | - | - | - | 52.33 | - |
RetinaNet [46] | ✗ | 51.20 | 73.79 | 62.50 | 61.57 | 41.17 | 69.06 | 42.02 | 53.46 | 64.50 | 52.75 | 51.04 | 33.75 | 43.77 | 38.51 | 28.71 | - | - | - | - | 44.72 | - |
RRPN [45] | ✔ | 61.80 | 70.40 | 66.10 | 69.80 | 43.94 | 72.18 | 35.90 | 55.46 | 77.16 | 57.85 | 58.77 | 37.45 | 55.68 | 48.17 | 39.94 | 77.88 | 55.12 | 25.22 | 40.29 | 53.57 | 52.14 |
R2CNN [43] | ✔ | 69.75 | 76.27 | 73.01 | 76.68 | 48.96 | 75.89 | 40.78 | 60.58 | 85.48 | 69.20 | 55.75 | 39.57 | 56.26 | 59.76 | 40.81 | 77.86 | 75.52 | 20.22 | 50.00 | 58.12 | 57.31 |
SCRDet [44] | ✔ | 71.79 | 77.70 | 74.75 | 80.06 | 46.56 | 83.66 | 36.75 | 61.76 | 85.97 | 72.32 | 58.82 | 40.56 | 57.22 | 54.74 | 38.13 | 89.29 | 76.57 | 16.88 | 45.68 | 58.25 | 57.83 |
RBox-CNN [47] | ✔ | 81.63 | 80.67 | 81.15 | 89.93 | 51.95 | 83.28 | 51.90 | 69.27 | 93.13 | 84.28 | 60.15 | 41.65 | 60.88 | 58.80 | 44.24 | 87.14 | 70.01 | 17.98 | 68.45 | 63.30 | 62.43 |
RetinaNet [46] | ✔ | 60.45 | 69.74 | 65.10 | 68.51 | 43.66 | 71.33 | 34.12 | 54.41 | 76.90 | 56.86 | 52.55 | 38.93 | 53.77 | 47.34 | 38.71 | 86.41 | 54.53 | 24.83 | 35.27 | 52.15 | 51.46 |
Coarse1 Label (L1) | Coarse2 Label (L2) | Fine Label (L3) | ||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Base Detector | +DHCR | S | A | mAP | CS | MS | CA | MA | mAP | CO | OT | MV | AC | SU | LW | SW | CA1 | CA2 | MA1 | MA2 | mAP† | mAP |
RRPN [43] | ✗ | 42.40 | 50.01 | 46.21 | 43.29 | 40.04 | 65.20 | 30.71 | 44.81 | 48.68 | 30.09 | 32.11 | 51.57 | 32.02 | 45.54 | 28.87 | - | - | - | - | 38.41 | - |
R2CNN [44] | ✗ | 50.32 | 54.51 | 52.42 | 50.55 | 42.75 | 64.73 | 40.98 | 49.75 | 68.90 | 36.84 | 37.76 | 53.25 | 38.57 | 46.56 | 30.50 | - | - | - | - | 44.63 | - |
SCRDet [45] | ✗ | 49.81 | 60.14 | 54.98 | 53.44 | 40.04 | 67.17 | 43.67 | 51.08 | 68.05 | 34.68 | 40.88 | 50.59 | 40.27 | 48.88 | 31.05 | - | - | - | - | 44.91 | - |
RBox-CNN [47] | ✗ | 55.75 | 70.24 | 63.00 | 60.05 | 43.99 | 80.43 | 45.07 | 57.39 | 85.45 | 45.12 | 45.46 | 55.11 | 41.05 | 53.21 | 33.32 | - | - | - | - | 51.25 | - |
RetinaNet [46] | ✗ | 42.23 | 51.78 | 47.01 | 45.27 | 41.12 | 66.85 | 30.85 | 46.02 | 50.66 | 31.11 | 33.57 | 50.99 | 37.76 | 45.05 | 29.05 | - | - | - | - | 39.74 | - |
RRPN [45] | ✔ | 44.26 | 55.41 | 49.84 | 48.97 | 40.25 | 65.43 | 31.78 | 46.61 | 59.12 | 33.40 | 36.99 | 56.93 | 43.16 | 47.72 | 30.29 | 71.69 | 41.47 | 37.63 | 6.43 | 43.94 | 42.26 |
R2CNN [43] | ✔ | 51.83 | 62.92 | 57.38 | 53.85 | 48.97 | 73.99 | 35.76 | 53.14 | 80.63 | 51.29 | 41.82 | 56.19 | 44.24 | 49.07 | 31.85 | 83.96 | 42.12 | 47.63 | 19.51 | 50.73 | 49.85 |
SCRDet [44] | ✔ | 55.43 | 64.23 | 59.83 | 58.70 | 44.70 | 75.23 | 34.14 | 53.19 | 81.66 | 52.76 | 44.76 | 55.71 | 46.12 | 50.88 | 33.45 | 80.89 | 35.93 | 53.25 | 20.81 | 52.19 | 50.57 |
RBox-CNN [47] | ✔ | 61.26 | 70.85 | 66.06 | 62.97 | 50.14 | 80.29 | 40.55 | 58.49 | 90.40 | 60.32 | 45.86 | 60.12 | 47.27 | 55.63 | 34.76 | 85.79 | 35.14 | 55.21 | 23.83 | 56.34 | 54.03 |
RetinaNet [46] | ✔ | 45.23 | 57.97 | 51.60 | 49.32 | 42.78 | 70.81 | 35.40 | 49.58 | 60.59 | 35.00 | 40.17 | 56.51 | 43.24 | 46.32 | 30.81 | 72.45 | 35.19 | 40.78 | 15.66 | 44.66 | 43.34 |
Raw Input Size (M) | Performance of RPN | Training Time | Inference Time | ||
---|---|---|---|---|---|
before Down-Scaling | TP Rate (%) | FN Rate (%) | mAP (%) | (h) | (s/img) |
3000 × 3000 | 51.14 | 48.86 | 31.09 | 218.00 | 45.46 |
5000 × 5000 | 36.52 | 63.48 | 26.68 | 208.07 | 40.42 |
7000 × 7000 | 35.47 | 64.53 | 24.58 | 195.35 | 38.76 |
9000 × 9000 | 29.04 | 70.96 | 21.08 | 176.12 | 31.33 |
RRPN [43] | R2CNN [44] | SCRDet [45] | RBox-CNN [47] | RetinaNet [46] | DHCR (Ours) |
---|---|---|---|---|---|
0.3134 | 0.3621 | 0.3977 | 0.4535 | 0.2799 | 0.4873 |
Author Contributions
Conceptualization, all authors; methodology, all authors; validation, all authors; formal analysis, all authors; investigation, S.-J.S., S.K. (Seyeob Kim) and Y.K.; resources, S.K. (Sungho Kim); data curation, S.-J.S. and S.K. (Seyeob Kim); writing-original draft preparation, S.-J.S.; writing-review and editing, S.K. (Seyeob Kim), Y.K. and S.K. (Sungho Kim); visualization, S.-J.S.; supervision, S.K. (Sungho Kim); project administration, S.K. (Sungho Kim). All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Conflicts of Interest
The authors declare no conflict of interest.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2020. This work is licensed under http://creativecommons.org/licenses/by/3.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Detecting objects such as aircraft and ships is a fundamental research area in remote sensing analytics. Owing to the prosperity and development of CNNs, many previous methodologies have been proposed for object detection within remote sensing images. Despite the advance, using the object detection datasets with a more complex structure, i.e., datasets with hierarchically multi-labeled objects, is limited to the existing detection models. Especially in remote sensing images, since objects are obtained from bird’s-eye view, the objects are captured with restricted visual features and not always guaranteed to be labeled up to fine categories. We propose a hierarchical multi-label object detection framework applicable to hierarchically partial-annotated datasets. In the framework, an object detection pipeline called Decoupled Hierarchical Classification Refinement (DHCR) fuses the results of two networks: (1) an object detection network with multiple classifiers, and (2) a hierarchical sibling classification network for supporting hierarchical multi-label classification. Our framework additionally introduces a region proposal method for efficient detection on vain areas of the remote sensing images, called clustering-guided cropping strategy. Thorough experiments validate the effectiveness of our framework on our own object detection datasets constructed with remote sensing images from WorldView-3 and SkySat satellites. Under our proposed framework, DHCR-based detections significantly improve the performance of respective baseline models and we achieve state-of-the-art results on the datasets.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer