Fusing DCN and BBAV for Remote Sensing Image Object Detection

Fusing DCN and BBAV for Remote Sensing Image Object Detection

Honghuan Chen, Keming Wang
DOI: 10.4018/IJCINI.335496
Article PDF Download
Open access articles are freely available for download

Abstract

At the oriented object detection in aerial remote sensing images, the perceptual field boundaries of ordinary convolutional kernels are often not parallel to the boundaries of the objects to be detected, affecting the model precision. Therefore, an object detection model (DCN-BBAV) that fuses deformable convolution networks (DCNs) and box boundary-aware vectors (BBAVs) is proposed. Firstly, a BBAV is used as the baseline, replacing the normal convolution kernels in the backbone network with deformable convolution kernels. Then, the spatial attention module (SAM) and channel attention mechanism (CAM) are used to enhance the feature extraction ability for a DCN. Finally, the dot product of the included angles of four adjacent vectors are added to the loss function of the rotation frame parameter, improving the regression precision of the boundary vector. The DCN-BBAV model demonstrates notable performance with a 77.30% mean average precision (mAP) on the DOTA dataset. Additionally, it outperforms other advanced rotating frame object detection methods, achieving impressive results of 90.52% mAP on VOC07 and 96.67% mAP on VOC12 for HRSC2016.
Article Preview
Top

Introduction

Aerial object detection is a key computer vision task (Ding et al., 2021; Wang et al., 2020; Hu et al., 2022) that has been getting increasing attention in recent years and plays a significant role in remote image understanding. Unlike general object detection, aerial object localization presents particularly tricky questions, including nonaxis-aligned objects in arbitrary directions (Ding et al., 2019; Han et al., 2021; Pan et al., 2020) and dense distributions in complex contexts (Guo et al., 2021; Yang et al., 2018, 2021). For example, aircraft object detection techniques are mainly interfered with by external factors, for instance, noise, weather, light intensity, shadows, and background (Xiaolin et al., 2021) in remote sensing images.

Mainstream methods usually treat aerial object detection as a question of rotating object localization (Han et al., 2020; Yang et al., 2020a). Among them, the angle-based direct orientation regression method is dominant in this research area, and it comes from general detectors (Lin et al., 2016; Li et al., 2022a; Lu et al., 2022; Zhang et al., 2022) with additional orientation parameters. While promising performance has been achieved, direct orientation prediction still suffers from a number of problems, including loss discontinuities and regression inconsistencies (Wang et al., 2022; Yang et al., 2020a, 2021; Yang & Yan 2022). The reasons are the bounded periodicity of the angular directions and the orientation definition of the rotating bounding box. Detectors based on orientation regression may not be able to accurately predict the orientation despite their attractive localization results.

To effectively address the aforementioned issues, the representation of airborne objects is revisited in order to prevent orientation estimation that is overly sensitive. Point sets are exceptionally capable of capturing important semantic features in conventional general-purpose detectors, such as RepPoints, as a fine-grained object representation (Yao et al., 2022). However, its basic transformation function can only generate upright-horizontal bounding boxes, which are unable to precisely calculate the orientation of airborne objects with precision. Additionally, RepPoints ignores a measure of the learned point's quality and merely regresses significant points. Poor performance for complex scenes and nonaxis-aligned objects with dense distribution may result from this in aerial images.

Complete Article List

Search this Journal:
Reset
Volume 18: 1 Issue (2024)
Volume 17: 1 Issue (2023)
Volume 16: 1 Issue (2022)
Volume 15: 4 Issues (2021)
Volume 14: 4 Issues (2020)
Volume 13: 4 Issues (2019)
Volume 12: 4 Issues (2018)
Volume 11: 4 Issues (2017)
Volume 10: 4 Issues (2016)
Volume 9: 4 Issues (2015)
Volume 8: 4 Issues (2014)
Volume 7: 4 Issues (2013)
Volume 6: 4 Issues (2012)
Volume 5: 4 Issues (2011)
Volume 4: 4 Issues (2010)
Volume 3: 4 Issues (2009)
Volume 2: 4 Issues (2008)
Volume 1: 4 Issues (2007)
View Complete Journal Contents Listing