Pretrained Detail Enhancement Framework for Remote Sensing Object Detection

Remote sensing object detection faces significant challenges due to the varying resolutions of images and the diverse shapes of objects. In this paper, we introduce an innovative approach to enhance object detection networks in remote sensing images by addressing these issues through a novel two-sta...

Full description

Saved in:
Bibliographic Details
Main Authors: Mo Zhou, Yue Zhou, Dawei Yang, Kai Song
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10829622/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1841542566862913536
author Mo Zhou
Yue Zhou
Dawei Yang
Kai Song
author_facet Mo Zhou
Yue Zhou
Dawei Yang
Kai Song
author_sort Mo Zhou
collection DOAJ
description Remote sensing object detection faces significant challenges due to the varying resolutions of images and the diverse shapes of objects. In this paper, we introduce an innovative approach to enhance object detection networks in remote sensing images by addressing these issues through a novel two-stage training process. Our key contributions include the development of a combined low-resolution pretraining and high-resolution fine-tuning strategy, which significantly improves the model’s generalization capability across different resolutions. We have designed a Pretrained Detail Enhancement and Restoration (PDER) framework that enhances feature extraction from low-resolution images, thereby boosting the baseline network’s ability to handle blurred and intricate objects. Additionally, we propose an Energy-Efficient Shape Adaptive Convolution (EESA) that integrates a depthwise separable mechanism with shape-adaptive feature maps, enabling efficient capture of complex object edges without substantial computational overhead. Compared to state-of-the-art methods such as YOLOv7, our model achieves an absolute average precision improvement of 2.34% and 2.57% on the DIOR and xView datasets, respectively. Extensive experiments on multiple remote sensing datasets demonstrate our approach’s effectiveness, achieving significant improvements in detection accuracy and computational efficiency, thus highlighting its potential for practical applications.
format Article
id doaj-art-8f7a91a577694ef49d57d577c290f9bb
institution Kabale University
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-8f7a91a577694ef49d57d577c290f9bb2025-01-14T00:02:12ZengIEEEIEEE Access2169-35362025-01-01136362637710.1109/ACCESS.2025.352618010829622Pretrained Detail Enhancement Framework for Remote Sensing Object DetectionMo Zhou0https://orcid.org/0009-0002-8926-3722Yue Zhou1Dawei Yang2https://orcid.org/0000-0003-0717-9579Kai Song3https://orcid.org/0000-0002-8735-1609School Information Science and Engineering, Shenyang Ligong University, Shenyang, Liaoning, ChinaSchool Information Science and Engineering, Shenyang Ligong University, Shenyang, Liaoning, ChinaSchool Information Science and Engineering, Shenyang Ligong University, Shenyang, Liaoning, ChinaSchool Information Science and Engineering, Shenyang Ligong University, Shenyang, Liaoning, ChinaRemote sensing object detection faces significant challenges due to the varying resolutions of images and the diverse shapes of objects. In this paper, we introduce an innovative approach to enhance object detection networks in remote sensing images by addressing these issues through a novel two-stage training process. Our key contributions include the development of a combined low-resolution pretraining and high-resolution fine-tuning strategy, which significantly improves the model’s generalization capability across different resolutions. We have designed a Pretrained Detail Enhancement and Restoration (PDER) framework that enhances feature extraction from low-resolution images, thereby boosting the baseline network’s ability to handle blurred and intricate objects. Additionally, we propose an Energy-Efficient Shape Adaptive Convolution (EESA) that integrates a depthwise separable mechanism with shape-adaptive feature maps, enabling efficient capture of complex object edges without substantial computational overhead. Compared to state-of-the-art methods such as YOLOv7, our model achieves an absolute average precision improvement of 2.34% and 2.57% on the DIOR and xView datasets, respectively. Extensive experiments on multiple remote sensing datasets demonstrate our approach’s effectiveness, achieving significant improvements in detection accuracy and computational efficiency, thus highlighting its potential for practical applications.https://ieeexplore.ieee.org/document/10829622/Remote sensingobject detectiontransfer learningsuper-resolution
spellingShingle Mo Zhou
Yue Zhou
Dawei Yang
Kai Song
Pretrained Detail Enhancement Framework for Remote Sensing Object Detection
IEEE Access
Remote sensing
object detection
transfer learning
super-resolution
title Pretrained Detail Enhancement Framework for Remote Sensing Object Detection
title_full Pretrained Detail Enhancement Framework for Remote Sensing Object Detection
title_fullStr Pretrained Detail Enhancement Framework for Remote Sensing Object Detection
title_full_unstemmed Pretrained Detail Enhancement Framework for Remote Sensing Object Detection
title_short Pretrained Detail Enhancement Framework for Remote Sensing Object Detection
title_sort pretrained detail enhancement framework for remote sensing object detection
topic Remote sensing
object detection
transfer learning
super-resolution
url https://ieeexplore.ieee.org/document/10829622/
work_keys_str_mv AT mozhou pretraineddetailenhancementframeworkforremotesensingobjectdetection
AT yuezhou pretraineddetailenhancementframeworkforremotesensingobjectdetection
AT daweiyang pretraineddetailenhancementframeworkforremotesensingobjectdetection
AT kaisong pretraineddetailenhancementframeworkforremotesensingobjectdetection