4.7 Article

Oriented Bounding Boxes for Small and Freely Rotated Objects

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TGRS.2021.3076050

Keywords

Object detection; Feature extraction; Proposals; Remote sensing; Marine vehicles; Task analysis; Satellites; Object detection; oriented bounding box; remote sensing; rotation-invariant; tiny objects

Funding

  1. Natural Sciences and Engineering Research Council (NSERC) Canada

Ask authors/readers for more resources

The proposed novel object detection method can handle freely rotated objects of arbitrary sizes, including tiny objects as small as 2 x 2 pixels. It uses a convolutional neural network (CNN) approach without external resources like anchor boxes, encodes location and orientation information by classification, and improves performance over existing state-of-the-art methods on data sets such as xView and DOTA.
A novel object detection method is presented that handles freely rotated objects of arbitrary sizes, including tiny objects as small as 2 x 2 pixels. Such tiny objects appear frequently in remotely sensed images, and present a challenge to recent object detection algorithms. More importantly, current object detection methods have been designed originally to accommodate axis-aligned bounding box detection, and therefore fail to accurately localize oriented boxes that best describe freely rotated objects. In contrast, the proposed convolutional neural network (CNN) -based approach uses potential pixel information at multiple scale levels without the need for any external resources, such as anchor boxes. The method encodes the precise location and orientation of features of the target objects at grid cell locations. Unlike existing methods that regress the bounding box location and dimension, the proposed method learns all the required information by classification, which has the added benefit of enabling oriented bounding box detection without any extra computation. It thus infers the bounding boxes only at inference time by finding the minimum surrounding box for every set of the same predicted class labels. Moreover, a rotation-invariant feature representation is applied to each scale, which imposes a regularization constraint to enforce covering the 360 degrees range of in-plane rotation of the training samples to share similar features. Evaluations on the xView and dataset for object detection in aerial images (DOTA) data sets show that the proposed method uniformly improves performance over existing state-of-the-art methods.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available