Back to articles
Volume: 32 | Article ID: art00002
A New Training Model for Object Detection in Aerial Images
  DOI :  10.2352/ISSN.2470-1173.2020.8.IMAWM-084  Published OnlineJanuary 2020

This paper presents a new training model for orientation invariant object detection in aerial images by extending a deep learning based RetinaNet which is a single-stage detector based on feature pyramid networks and focal loss for dense object detection. Unlike R3Det which applies feature refinement to handle rotating objects, we proposed further improvement to cope with the densely arranged and class imbalance problems in aerial imaging, on three aspects: 1) All training images are traversed in each iteration instead of only one image in each iteration in order to cover all possibilities; 2) The learning rate is reduced if losses are not reduced; and 3) The learning rate is reduced if losses are not changed. The proposed method was calibrated and validated by comprehensive for performance evaluation and benchmarking. The experiment results demonstrate the significant improvement in comparison with R3Dec approach on the same data set. In addition to the well-known public data set DOTA for benchmarking, a new data set is also established by considering the balance between the training set and testing set. The map of losses which dropped down smoothly without jitter and overfitting also illustrates the advantages of the proposed newmodel.

Subject Areas :
Views 77
Downloads 5
 articleview.views 77
 articleview.downloads 5
  Cite this article 

Geng YANG, Yu Geng, Qin LI, Jane YOU, Mingpeng Cai, "A New Training Model for Object Detection in Aerial Imagesin Proc. IS&T Int’l. Symp. on Electronic Imaging: Imaging and Multimedia Analytics in a Web and Mobile World,  2020,  pp 84-1 - 84-5,

 Copy citation
  Copyright statement 
Copyright © Society for Imaging Science and Technology 2020
Electronic Imaging
Society for Imaging Science and Technology
7003 Kilworth Lane, Springfield, VA 22151 USA