Multi-Level Foreground Prompt for Incremental Object Detection

In the study of incremental object detection, knowledge distillation and data replay are effective methods to mitigate catastrophic forgetting. However, current research on single-stage detectors is limited, single-stage detector outputs often contain excessive negative sample information, and direc...

Full description

Saved in:
Bibliographic Details
Main Authors: Jianwen Mo, Ronghua Zou, Hua Yuan
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10819438/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1841550814520279040
author Jianwen Mo
Ronghua Zou
Hua Yuan
author_facet Jianwen Mo
Ronghua Zou
Hua Yuan
author_sort Jianwen Mo
collection DOAJ
description In the study of incremental object detection, knowledge distillation and data replay are effective methods to mitigate catastrophic forgetting. However, current research on single-stage detectors is limited, single-stage detector outputs often contain excessive negative sample information, and direct application of knowledge distillation to them is ineffective. To address this, this paper proposes a multi-level foreground prompt incremental learning algorithm for single-stage detectors like YOLO, including foreground prompts at the image level, feature map level, and knowledge level. First, to obtain fewer but more representative replay samples, representative images with a high number of old class foregrounds are selected by maximizing sample structure distance, providing direct foreground information at the image level. Second, the feature map output by the teacher model is used as a feature-level prompt, with a feature distillation loss guiding the student model to encode new class foreground information in less significant channels of the old feature map, reducing interference. Lastly, the teacher model’s inference output serves as a knowledge-level prompt, and the adaptive select object method is introduced to avoid foreground conflict in traditional knowledge distillation, enhancing the model’s plasticity by selectively merging foreground information. Extensive experiments on PASCAL VOC and MS COCO datasets demonstrate that this approach significantly improves model plasticity while maintaining stability.
format Article
id doaj-art-fdd2ddda95ce4f55886265a4b2a42e14
institution Kabale University
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-fdd2ddda95ce4f55886265a4b2a42e142025-01-10T00:01:32ZengIEEEIEEE Access2169-35362025-01-01134048406610.1109/ACCESS.2024.352453910819438Multi-Level Foreground Prompt for Incremental Object DetectionJianwen Mo0https://orcid.org/0000-0002-1729-1284Ronghua Zou1https://orcid.org/0009-0008-0050-3181Hua Yuan2https://orcid.org/0009-0001-8592-9884School of Information and Communication, Guilin University of Electronic Technology, Guilin, ChinaSchool of Information and Communication, Guilin University of Electronic Technology, Guilin, ChinaSchool of Information and Communication, Guilin University of Electronic Technology, Guilin, ChinaIn the study of incremental object detection, knowledge distillation and data replay are effective methods to mitigate catastrophic forgetting. However, current research on single-stage detectors is limited, single-stage detector outputs often contain excessive negative sample information, and direct application of knowledge distillation to them is ineffective. To address this, this paper proposes a multi-level foreground prompt incremental learning algorithm for single-stage detectors like YOLO, including foreground prompts at the image level, feature map level, and knowledge level. First, to obtain fewer but more representative replay samples, representative images with a high number of old class foregrounds are selected by maximizing sample structure distance, providing direct foreground information at the image level. Second, the feature map output by the teacher model is used as a feature-level prompt, with a feature distillation loss guiding the student model to encode new class foreground information in less significant channels of the old feature map, reducing interference. Lastly, the teacher model’s inference output serves as a knowledge-level prompt, and the adaptive select object method is introduced to avoid foreground conflict in traditional knowledge distillation, enhancing the model’s plasticity by selectively merging foreground information. Extensive experiments on PASCAL VOC and MS COCO datasets demonstrate that this approach significantly improves model plasticity while maintaining stability.https://ieeexplore.ieee.org/document/10819438/Catastrophic forgettingincremental learningknowledge distillationobject detection
spellingShingle Jianwen Mo
Ronghua Zou
Hua Yuan
Multi-Level Foreground Prompt for Incremental Object Detection
IEEE Access
Catastrophic forgetting
incremental learning
knowledge distillation
object detection
title Multi-Level Foreground Prompt for Incremental Object Detection
title_full Multi-Level Foreground Prompt for Incremental Object Detection
title_fullStr Multi-Level Foreground Prompt for Incremental Object Detection
title_full_unstemmed Multi-Level Foreground Prompt for Incremental Object Detection
title_short Multi-Level Foreground Prompt for Incremental Object Detection
title_sort multi level foreground prompt for incremental object detection
topic Catastrophic forgetting
incremental learning
knowledge distillation
object detection
url https://ieeexplore.ieee.org/document/10819438/
work_keys_str_mv AT jianwenmo multilevelforegroundpromptforincrementalobjectdetection
AT ronghuazou multilevelforegroundpromptforincrementalobjectdetection
AT huayuan multilevelforegroundpromptforincrementalobjectdetection