Professional Documents
Culture Documents
1
detection steps be integrated in an unified network, which troduced which enforces the samples and their rotated ver-
achieves the best detected result at that time. Anchor boxes sions share the similar features [5]. Unlike these methods
with multiple scales and aspect ratios are of the essence which try to eliminate the effect of rotation on the feature
in Faster R-CNN, which locate the position of candidate level, we prefer to make the rotation information useful for
objects. In Fast/Faster R-CNN, the network can be parti- feature extraction so that the detection results involve the
tioned into a fully convolutional subnetwork which is in- angle information of the objects. Therefore, the detection
dependent of proposals and shares the computation, and a results is rotatable, whereas the performance of the detector
per-proposal subnetwork without sharing the computation. is rotation invariant.
It is noticed that the per-proposal subnetwork is inefficient, Zhou et al. proposed a network named ORN [27] which
and a region-based fully convolutional network (R-FCN) is shows similar character with our method. ORN is used to
proposed in [6] to remedy that issue by using FCN to share classify images while extracting the orientation angle of the
almost all computation on the entire image. By adding a whole image. However, ORN can not be applied straight-
branch to predict the mask of objects upon Faster R-CNN, forward as a detection network, which needs to detect ori-
Mask R-CNN can simultaneously detect the object and gen- entation locally on each object. In our method, the angle
erate the segmentation mask of the object [12]. There is estimation is associated to plenty of prior boxes, so that the
another kind of DL-based object detection approach that rotation of an object can be realized by the corresponding
does not rely on region proposals, such as you only look prior boxes, while other prior boxes are still available for
once detector (YOLO) [18] and single shot multibox detec- other objects. Additionally, our method effectively sepa-
tor (SSD) [16]. We refer to YOLO and SSD as box-based rates the object proposals with its background pixels, so the
methods since they generate several boxes for detecting ob- angle estimation can concentrate on the object without the
jects in the image according to certain rules. These boxes interfere of background.
are called as prior boxes and objects are supposed to locate The next section explains why rotatable bounding box
in or near certain prior box. SSD attempts to use pyrami- (RBox) is better than BBox for rotation invariant detection.
dal feature hierarchy computed from the convolutional net, In section 3 we discuss how RBox takes place of BBox to
however without reusing the higher resolution maps in the form the newly designed detection method, DRBox. DR-
pyramidal feature hierarchy. The feature pyramid network Box is tested on ship, airplane and vehicle detection in re-
(FPN) [15] better utilizes the pyramidal feature hierarchy mote sensing images, exhibiting definitely superiority com-
through introducing lateral connections which merge high- pared with traditional BBox based methods.
level semantic feature maps with coarse resolution and low-
level semantic feature maps with refined resolution. 2. Rotatable bounding box
Many recent machine learning based methods have been Most object Detection methods use bounding box
used for remote sensing object detection. Many machine (BBox) to locate target objects in images. A BBox is a
learning methods extract the candidate object feature, such rectangle parameterized by four variables: the center point
as histogram of oriented gradients (HOG) [23, 4], tex- position (two variables), the width, and the height. BBox
ture [3], bag-of-words (BoW) [22], regional covariance de- meets difficulties locating objects with different orientation
scriptor (RCD) [1], and interest point [25], followed by angles. In this situation, BBox cannot provide the exact
certain classifier, for example, sparse representation classi- sizes of objects, and is very difficult to distinguish dense
fier (SRC) [4, 3] and support vector machine (SVM) [1]. objects. Table 1 lists the three main disadvantages of BBox
As the development of deep learning technique, DNNs and their corresponding examples. As shown in this table,
have been successfully employed to solve object detection a ship target in remote sensing image is aligned to BBoxes
problem in remote sensing images. A CNN based pio- with different sizes and aspect ratios because of rotation. As
neering study upon vehicle detection in satellite images the result, the width and height of BBox have no relation-
is presented in[2], followed by many approaches that fo- ship with the physical size of the rotated target. Another sig-
cus on different types of geospatial object detection prob- nificant disadvantage is the discrimination between object
lems [14, 20, 21, 5, 26, 17]. As another effective fea- and background. In the given example, about 60% region
ture extraction tool, deep Boltzmann machine is also used inside the BBox belongs to background pixels when the ori-
as feature extractor in geo-spatial object detection prob- entation angle of the ship is near 45 degrees. The situation
lems [11, 7]. becomes more difficult when target objects are distributed
To make the approach insensitive to objects in-plane ro- dense, in which case the objects are hard to be separated by
tation, some efforts are made either adjusting the orienta- BBoxes.
tion, or trying to extract rotation insensitive features. For In this article, RBox is defined to overcome the above
example, to deal with the rotation variation of geo-spatial difficulties. RBox is a rectangle with a angle parameter to
objects issue, a rotation invariant regularization term is in- define its orientation. An RBox needs five parameters to de-
2
Disadvantages of the traditional bounding box which is defined as following:
1. The size 2. Object and 3. Dense ob-
and aspect ratios background pix- jects are difficult area  ∩ B
do not reflect the els are not effec- to be separated. ArIoU (A, B) = cos (θA − θB ) , (2)
area  ∪ B
real shape of the tively separated.
target object. or
area  ∩ B
ArIoU180 (A, B) = |cos (θA − θB )| , (3)
area  ∪ B
3
5HVXOW apply hard negative mining to decrease the number of the
'HFRGLQJ 106
negative samples.
2XWSXWUHVXOW
The objective loss function of DRBox is extended from
SSD objective loss function by adding the angle related
3UHGLFWLRQOD\HU K JURXSVIRU
term. The overall objective loss function is as following:
/RFDWLRQ HDFKSULRU
6L]H 5%R[ 1
$QJOH 0XOWL$QJOH3ULRU5%R[HV
L (x, c, l, g) = (Lconf (c) + Lrbox (x, l, g)) (4)
&RQILGHQFH 灤FRQY N
瀽
瀽澳 瀽
where N is the number of matched prior RBoxes. The con-
瀽 瀽
)HDWXUH fidence loss Lconf (c) is a two class softmax loss over all se-
([WUDFWLRQ «« 瀽 lected positive and negative samples, where c is the two-
/D\HUV
dimension confidence vector. The RBox regression loss
Lrbox (x, l, g) is similar to SSD and Faster R-CNN, where we
calculate the smooth L1 loss between the predicted RBox l
,QSXWLPDJH and the ground truth RBox g:
Lrbox (x, l, g)
Figure 1: The networks structure of DRBox. The networks X X X
structure of DRBox is similar with other box based methods = xij smoothL1 ˆlim − ĝjm
except for the use of multi-angle prior RBoxes. DRBox i∈P os j m∈{cx,cy,w,h,a}
searches for objects using sliding and rotating prior RBoxes (5)
on input image and then output locations of objects besides
where xij ∈ {1, 0} is an indicator for matching the i-th
with their orientation angles.
prior RBox to the j-th ground truth RBox. ˆl and ĝ are de-
fined as following, which are the offsets of the parameters in
Multi-angle prior RBox plays an important role in DR- l and g with their corresponding prior RBox p, respectively:
Box. The convolutional structure ensures that prior boxes
can move over different locations to search for target ob- t̂cx = (tcx − pcx )/pw , t̂cy = (tcy − pcy ) ph ; (6a)
jects. On each location, the prior RBoxes rotate at a series of
angles to generate multi-angle predictions, which is the key t̂w = log (tw /pw ) , t̂h = log th ph ; (6b)
difference between DRBox and other bounding box based
methods. The aspect ratio used in detection is fixed accord- t̂a = tan (ta − pa ) . (6c)
ing to the object type, which decreased the total number Equations 6a , 6b and 6c are the location regression
of prior boxes. By the multi-angle prior RBoxes strategy, terms, the size regression terms and the angle regression
the network is trained to treat the detection task as a series term, respectively. The angle regression term applies tan-
of sub-tasks. Each sub-task focuses on one narrow angle gent function to adapt to the periodicity of the angle param-
range, therefore decreases the difficulty caused by rotation eter. The minimization of the angle regression term ensures
of objects. that the correct angle is learned during training.
3.2. Training 3.3. Complement details
The training of DRBox is extended from SSD train- PYRAMID INPUT: DRBox applies pyramid input strat-
ing procedure [16] to involve angle estimation. During egy that the original image is rescaled into different res-
training, each ground truth RBox is assigned with several olutions, and separated into overlapped 300 × 300 sub-
prior RBoxes according to their ArIoU. ArIoU is a non- images. The DRBox network is applied to each sub-image
commutative function, which means ArIoU(A, B) is dif- and the network only detects targets with proper size. Non-
ferent from ArIoU(B, A). A prior RBox P is assigned to maximum suppression is applied on the detection results of
a ground truth RBox G when ArIoU(P, G) > 0.5. After the whole image, which suppresses repeated predictions not
the assignment, the matched prior RBoxes are considered only within a sub-image, but also crossing overlap areas of
as positive samples and are responsible for generating the different sub-images. The pyramid input strategy helps the
losses of location and angle regression. The use of ArIoU detection network to share features between large and small
helps the training process to select prior RBox with proper objects. Besides, the satellite image used in this article is
angle as positive samples, so the angle information of the often very large, so the division process and non-maximum
objects can be roughly learned during training. After the suppression across sub-images helps to detect objects in
matching step, most of the prior RBoxes are negative. We very large images.
4
CONVOLUTION ARCHITECTURE: DRBox uses angle of the object. A Matlab tool is developed to label the
truncated VGG-net for detection. All the full-connective data with RBoxes.
layers, convolution layers and pooling layers after layer
conv4 3 are removed. Then, a 3 × 3 convolution layer is 4.2. Benchmark
added after layer conv4 3. The receptive field of DRBox In this section, we compare the performance of DRBox
is 108 pixels 108 pixels, so any targets larger than this with the detectors that use BBox. SSD and Faster R-CNN
scope cannot be detected. Besides, the feature map of layer are used as benchmarks of BBox based methods. All the de-
conv4 3 is 38 × 38, so the targets closer than 8 pixels may tectors use the same convolution architecture and the same
be missed. training data argumentation strategy. The prior boxes used
PRIOR RBOX SETTINGS: In this article, three DRBox in SSD and the anchor boxes used in Faster R-CNN are op-
networks are trained separately for vehicle detection, ship timized for the datasets. All other hyper parameters are op-
detection and airplane detection, respectively. The scale timized for the dataset, too.
and input resolution settings jointly ensure that the areas of
the prior RBoxes cover the sizes of the objects sufficiently, 4.3. Detection results
thus the objects of different sizes can be effectively cap-
Figure 2 show ships, vehicles and airplanes detected
tured. In ship detection, it is hard to distinguish the head
by using RBox (DRBox) and BBox (SSD). The predicted
and tail of the target. In this case, the angles of the ground
bounding boxes that matches the ground truth bounding
truth RBoxes and multi-angle prior RBoxes varies from 0
boxes with IoU > 0.5 are plotted in green color, while the
degree to 180 degrees. In detail, ship objects are detected
false positive predictions and false negative predictions are
with prior RBoxes of 20 × 8, 40 × 14, 60 × 17, 80 × 20,
plotted in yellow and red color, respectively. Our method is
100 × 25 pixels in size and 0:30:150 degrees in angle; vehi-
better in the given scenes. DRBox successfully detects most
cle objects are detected with prior RBoxes of 25 × 9 pixels
of the ships on both the port region and open water region,
in size and 0:30:330 degrees in angle; airplane objects are
while SSD almost fails on detection of nearby ships. The
detected with prior RBoxes of 50 × 50, 70 × 70 pixels in
detection results for vehicles and airplanes also show that
size and 0:30:330 degrees in angle. The total number of
SSD generates more false alarms and false dismissals than
prior boxes per image is 43320, 17328 and 34656 for ship,
does DRBox.
vehicle and airplane detection, respectively.
More results of DRBox are shown in 3. Ship detection
DRBox reaches 70-80 fps on NVIDIA GTX 1080Ti and
in port region is more challengeable than in open water re-
Intel Core i7. The input pyramid strategy produce no more
gion, whereas DRBox works well on both situations. The
than 4/3 times time cost. Considering 1/3 overlapped be-
vehicles are very difficult to detect due to small size and
tween sub-images, DRBox reaches processing a speed of
complex backgrounds. In our dataset, each car is around
1600 × 1600 pixels2 per second. The speed of SSD and
20 pixels in length and 9 pixels in width. Fortunately, we
Faster R-CNN are 70 fps and 20 fps on our dataset using
find that DRBox successfully finds vehicles that hided in
the same convolution network architecture.
the shadows of tall buildings, or parked very close to each
other. We also find that DRBox can not only output the
4. Experiments and results locates of the cars, but also predict the head direction of
4.1. Dataset each car, which is even a challenge task for human beings.
The estimated direction of cars on road matches the prior
We apply our method on object detection of satellite im- knowledge that traffic always keeps to the right of the road.
ages. We have not found any open source dataset on this Airplanes with different sizes are succesfully detected, in-
problem, so we build one using the GoogleEarth images. cluding an airplane that is under repairing.
The dataset includes three categories of objects: vehicles, Figure 4 shows precision-recall (P-R) curves of DRBox,
ships and airplanes. The vehicles are collected from ur- SSD and Faster R-CNN. The recall ratio evaluates the abil-
ban area in Beijing, China. The ships are collected near the ity of finding more targets in the image, while the preci-
wharfs and ports besides the Changjiang River, the Zhujiang sion evaluates the quality of predicting only the right object
River and the East China Sea. The airplanes are collected rather than containing many false alarms. The P-R curve of
from the images of 15 airports in China and America. The SSD and Faster R-CNN are always below the P-R curve of
dataset recently includes about 12000 vehicles, 3000 ships DRBox. DRBox has the best performance in this test. We
and 2000 airplanes and is still under expansion. About 2000 further show BEP (Break-Even Point), AP (Average Preci-
vehicles, 1000 ships and 500 airplanes are taken out as test- sion) and mAP (mean Average Precision) of each method
ing dataset and others are used for training. in Table 2. BEP is the point on P-R curve where precision
Each object in the images are marked with a RBox, equals recall, AP is the area below P-R curve, and mAP
which indicates not only the location and size, but also the is the mean value of the APs on all object detection tasks.
5
Method Dataset BEP(%) AP(%) mAP(%) Method Dataset STD AP(%) STD AS(%)
Ship 79.20 82.29 Ship 5.51 13.76
Faster Faster
Vehicle 71.60 75.55 85.63 Vehicle 7.21 7.82
R-CNN R-CNN
Airplane 98.07 99.06 Airplane 0.28 0.17
Ship 82.72 82.89 Ship 5.97 13.14
SSD Vehicle 83.13 87.59 89.68 SSD Vehicle 0.90 3.20
Airplane 97.74 98.56 Airplane 0.80 0.51
Ship 94.62 94.06 Ship 0.88 1.56
DRBox Vehicle 86.14 89.07 94.13 DRBox Vehicle 0.72 2.06
Airplane 98.62 99.28 Airplane 0.51 0.41
Table 2: BEP, AP and mAP of Faster R-CNN, SSD and Table 3: Quantitative evaluation of robustness against rota-
DRBox. DRBox outperforms the other two methods on all tion for Faster R-CNN, SSD and DRBox. STD AP eval-
indexes listed in this table. uates robustness against rotation of input images. STD R
evaluates robustness against rotation of objects. DRBox
DRBox is always the best on all the indexes compared with outperforms the other two methods except that Faster R-
SSD and Faster R-CNN. CNN is slightly more robust to rotation of airplane targets.
4.4. Comparison of the robustness against rotation angles, where the results of DRBox and SSD are plotted
in shallow red and shallow blue colors, respectively. The
The robustness against rotation of a method involves two curves generated by DRBox are more concentrated, which
aspects: robustness against rotation of input images and ro- indicates that DRBox is more robust against rotation of in-
bustness against rotation of objects. Robustness against ro- put images compared with SSD.
tation of input images means that a detection method should We calculate the recall ratio of targets in different angle
output the same results when the input image rotates arbi- scopes, respectively. Figure 6 show the results. Each angle
trarily. We define STD AP to quantify this ability. STD AP scope corresponds to one curve which shows the relation-
is the standard deviation of AP when the angle of the in- ship between the softmax threshold and recall ratio. The
put image changes. STD AP is estimated by rotating the performance of DRBox approximately remains the same for
test images each 10 degrees, then calculate the AP values each angle scopes, whereas the performance of SSD shows
of the detection results, respectively, and the standard de- strong instability when the angle of objects changes.
viation of the APs. Robustness against rotation of objects
means that the same object should always be successfully 5. Conclusions
detected when its orientation changes. We define STD AS
to quantify this ability. STD AS is the standard deviation Robustness to rotation is very important on detection
of the average score for objects in different orientation an- tasks of arbitrarily orientated objects. Existing detection al-
gles, which is estimated by dividing the objects in testing gorithms uses bounding box to locate objects, which is a ro-
dataset into different groups according to their angles, then tation variant structure. In this article, we replace the tradi-
calculate average score (softmax threshold) of each group, tional bounding box with RBox and reconstruct deep CNN
respectively, and the standard deviation over all groups. The based detection frameworks with this new structure. The
two robustness evaluation methods are interrelated, except proposed detector, which is called DRBox, is rotation in-
that STD AP has a bias towards the robustness on rotaton variant due to its ability of estimating the orientation angles
of backgrounds while STD AS has a bias towards the ro- of objects. DRBox outperforms Faster R-CNN and SSD on
bustness on rotation of objects. The smaller STD AP and object detection of satellite images.
STD AS value indicate that the detection method is more DRBox is designed as a box-based method, whereas it
robust against rotation. is also possible to apply RBox into proposal based detec-
STD AP and STD AS values of DRBox, SSD and Faster tion frameworks, e.g. R-FCN or Faster R-CNN. Training
R-CNN are shown in Table 3. All the three methods are ro- with RBox enforces the network to learn multi-scale local
bust against rotation on airplane detection. However, Faster orientation information of the input image. We are look-
R-CNN is relatively not robust against rotation on ship and ing forward of this interesting property to be used in other
vehicle detection, SSD is not robust against rotation on ship orientation sensitive tasks.
detection. DRBox remains good scores in all tasks.
Further comparison between DRBox and SSD are
demonstrated in Figure 5 and Figure 6. Figure 5 shows P-
R recalls of the same input image but rotated to different
6
Figure 2: Detection results using RBox and traditional bounding box. The first row are results of DRBox; the second row
are results of SSD. Columns 1-2, columns 3-4 and column 5 are results of ship detection, vehicle detection and airplane
detection, respectively. DRBox performs better in the given examples.
Figure 3: Detection results of DRBox. Ship detection results are shown in the first row, where DRBox works well in both
open water region and port region. Vehicle detection results are shown in the second row, where DRBox successfully finds
cars hidden in complex backgrounds and correctly indicates their head directions. Airplane detection results are shown in the
third row, including an airplane that is under repairing.
7
1 1
0.9 0.98
Precision
Precision
Precision
0.9 0.8 0.96
SSD SSD SSD
0.7 0.94
Faster Faster Faster
0.8 R-CNN 0.6 R-CNN R-CNN
0.92
DRBox DRBox DRBox
0.5
0 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.8 0.9 1
Recall Recall Recall
(a) (b) (c)
Figure 4: Precision-recall curves of (a) ship detection results, (b) vehicle detection results and (c) airplane detection results.
The performance of DRBox is the best in each detection task.
'5%R[
'5%R[
'5%R[ 66'
66' 66'
3UHFLVLRQ
3UHFLVLRQ
3UHFLVLRQ
5HFDOO 5HFDOO 5HFDOO
(a) (b) (c)
Figure 5: Precision-recall curves when input images are rotated on multiple angles in (a) ship detection task, (b) vehicle
detection task and (c) airplane detection task. The curves generated by SSD are in blue color, and the curves generated by
DRBox are in red color. The curves generated by DRBox are more concentrated, which indicates that DRBox is more robust
to rotation of input images.
5HFDOO
5HFDOO
5HFDOO
5HFDOO
5HFDOO
8
References electronics and Bioinformatics (ISBB), 2015 International
Symposium on, pages 184–187. IEEE, 2015.
[1] X. Chen, R.-X. Gong, L.-L. Xie, S. Xiang, C.-L. Liu, and C.-
[15] T.-Y. Lin, P. Dollár, R. Girshick, K. He, B. Hariharan, and
H. Pan. Building regional covariance descriptors for vehicle
S. Belongie. Feature pyramid networks for object detection.
detection. IEEE Geoscience and Remote Sensing Letters,
arXiv preprint arXiv:1612.03144, 2016.
14(4):524–528, 2017.
[16] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-
[2] X. Chen, S. Xiang, C.-L. Liu, and C.-H. Pan. Vehicle de- Y. Fu, and A. C. Berg. Ssd: Single shot multibox detector.
tection in satellite images by hybrid deep convolutional neu- In European conference on computer vision, pages 21–37.
ral networks. IEEE Geoscience and remote sensing letters, Springer, 2016.
11(10):1797–1801, 2014.
[17] Y. Long, Y. Gong, Z. Xiao, and Q. Liu. Accurate object lo-
[3] Z. Chen, C. Wang, H. Luo, H. Wang, Y. Chen, C. Wen, Y. Yu, calization in remote sensing images based on convolutional
L. Cao, and J. Li. Vehicle detection in high-resolution aerial neural networks. IEEE Transactions on Geoscience and Re-
images based on fast sparse representation classification and mote Sensing, 55(5):2486–2498, 2017.
multiorder feature. IEEE Transactions on Intelligent Trans-
[18] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi. You
portation Systems, 17(8):2296–2309, 2016.
only look once: Unified, real-time object detection. In Pro-
[4] Z. Chen, C. Wang, C. Wen, X. Teng, Y. Chen, H. Guan, ceedings of the IEEE Conference on Computer Vision and
H. Luo, L. Cao, and J. Li. Vehicle detection in high- Pattern Recognition, pages 779–788, 2016.
resolution aerial images via sparse representation and super- [19] S. Ren, K. He, R. Girshick, and J. Sun. Faster r-cnn: Towards
pixels. IEEE Transactions on Geoscience and Remote Sens- real-time object detection with region proposal networks. In
ing, 54(1):103–116, 2016. Advances in neural information processing systems, pages
[5] G. Cheng, P. Zhou, and J. Han. Learning rotation-invariant 91–99, 2015.
convolutional neural networks for object detection in vhr op- [20] I. Ševo and A. Avramović. Convolutional neural network
tical remote sensing images. IEEE Transactions on Geo- based automatic object detection on aerial images. IEEE
science and Remote Sensing, 54(12):7405–7415, 2016. Geoscience and Remote Sensing Letters, 13(5):740–744,
[6] J. Dai, Y. Li, K. He, and J. Sun. R-fcn: Object detection 2016.
via region-based fully convolutional networks. In Advances [21] L. W. Sommer, T. Schuchert, and J. Beyerer. Fast deep vehi-
in neural information processing systems, pages 379–387, cle detection in aerial images. In Applications of Computer
2016. Vision (WACV), 2017 IEEE Winter Conference on, pages
[7] W. Diao, X. Sun, X. Zheng, F. Dou, H. Wang, and K. Fu. 311–319. IEEE, 2017.
Efficient saliency-based object detection in remote sensing [22] H. Sun, X. Sun, H. Wang, Y. Li, and X. Li. Automatic tar-
images using deep belief networks. IEEE Geoscience and get detection in high-resolution remote sensing images using
Remote Sensing Letters, 13(2):137–141, 2016. spatial sparse coding bag-of-words model. IEEE Geoscience
[8] P. F. Felzenszwalb, R. B. Girshick, D. McAllester, and D. Ra- and Remote Sensing Letters, 9(1):109–113, 2012.
manan. Object detection with discriminatively trained part- [23] S. Tuermer, F. Kurz, P. Reinartz, and U. Stilla. Airborne
based models. IEEE transactions on pattern analysis and vehicle detection in dense urban areas using hog features and
machine intelligence, 32(9):1627–1645, 2010. disparity maps. IEEE Journal of Selected Topics in Applied
[9] R. Girshick. Fast r-cnn. In Proceedings of the IEEE inter- Earth Observations and Remote Sensing, 6(6):2327–2337,
national conference on computer vision, pages 1440–1448, 2013.
2015. [24] J. R. Uijlings, K. E. Van De Sande, T. Gevers, and A. W.
[10] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich fea- Smeulders. Selective search for object recognition. Interna-
ture hierarchies for accurate object detection and semantic tional journal of computer vision, 104(2):154–171, 2013.
segmentation. In Proceedings of the IEEE conference on [25] L. Wan, L. Zheng, H. Huo, and T. Fang. Affine invariant de-
computer vision and pattern recognition, pages 580–587, scription and large-margin dimensionality reduction for tar-
2014. get detection in optical remote sensing images. IEEE Geo-
[11] J. Han, D. Zhang, G. Cheng, L. Guo, and J. Ren. Ob- science and Remote Sensing Letters, 2017.
ject detection in optical remote sensing images based on [26] F. Zhang, B. Du, L. Zhang, and M. Xu. Weakly supervised
weakly supervised learning and high-level feature learn- learning based on coupled convolutional neural networks for
ing. IEEE Transactions on Geoscience and Remote Sensing, aircraft detection. IEEE Transactions on Geoscience and Re-
53(6):3325–3337, 2015. mote Sensing, 54(9):5553–5563, 2016.
[12] K. He, G. Gkioxari, P. Dollár, and R. Girshick. Mask r-cnn. [27] Y. Zhou, Q. Ye, Q. Qiu, and J. Jiao. Oriented response net-
arXiv preprint arXiv:1703.06870, 2017. works. In Proceedings of the IEEE Conference on Computer
[13] K. He, X. Zhang, S. Ren, and J. Sun. Spatial pyramid pooling Vision and Pattern Recognition, July 2017.
in deep convolutional networks for visual recognition. In [28] C. L. Zitnick and P. Dollár. Edge boxes: Locating object
European Conference on Computer Vision, pages 346–361. proposals from edges. In European Conference on Computer
Springer, 2014. Vision, pages 391–405. Springer, 2014.
[14] Q. Jiang, L. Cao, M. Cheng, C. Wang, and J. Li. Deep neural
networks-based vehicle detection in satellite images. In Bio-