EDADet: Encoder-Decoder Domain Augmented Alignment Detector for Tiny Objects in Remote Sensing Images

In recent years, deep learning has shown great potential in object detection applications, but it is still difficult to accurately detect tiny objects with an area proportion of less than 1% in remote sensing images. Most existing studies focus on designing complex networks to learn discriminative f...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on geoscience and remote sensing Vol. 63; pp. 1 - 15
Main Authors: Tao, Wenguang, Wang, Xiaotian, Yan, Tian, Bi, Haixia, Yan, Jie
Format: Journal Article
Language:English
Published: New York IEEE 01.01.2025
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects:
ISSN:0196-2892, 1558-0644
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In recent years, deep learning has shown great potential in object detection applications, but it is still difficult to accurately detect tiny objects with an area proportion of less than 1% in remote sensing images. Most existing studies focus on designing complex networks to learn discriminative features of tiny objects, usually resulting in a heavy computational burden. In contrast, this article proposes an accurate and efficient single-stage detector called EDADet for tiny objects. First, domain conversion technology is used to realize cross-domain multimodal data fusion based on single-modal data input. Then, a tiny object-aware backbone is designed to extract features at different scales. Next, an encoder-decoder feature fusion (EDFF) structure is devised to achieve efficient cross-scale propagation of semantic information. Finally, a center-assist loss and an alignment self-supervised loss are adopted to alleviate the position sensitivity issue and drift of tiny objects. A series of experiments on the AI-TODv2 dataset demonstrate the effectiveness and practicality of our EDADet. It achieves state-of-the-art (SOTA) performance and surpasses the second-best method by 9.65% in AP50 and 4.86% in mAP.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:0196-2892
1558-0644
DOI:10.1109/TGRS.2024.3510948