Multi-Level Foreground Prompt for Incremental Object Detection
In the study of incremental object detection, knowledge distillation and data replay are effective methods to mitigate catastrophic forgetting. However, current research on single-stage detectors is limited, single-stage detector outputs often contain excessive negative sample information, and direc...
Saved in:
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2025-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10819438/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1841550814520279040 |
---|---|
author | Jianwen Mo Ronghua Zou Hua Yuan |
author_facet | Jianwen Mo Ronghua Zou Hua Yuan |
author_sort | Jianwen Mo |
collection | DOAJ |
description | In the study of incremental object detection, knowledge distillation and data replay are effective methods to mitigate catastrophic forgetting. However, current research on single-stage detectors is limited, single-stage detector outputs often contain excessive negative sample information, and direct application of knowledge distillation to them is ineffective. To address this, this paper proposes a multi-level foreground prompt incremental learning algorithm for single-stage detectors like YOLO, including foreground prompts at the image level, feature map level, and knowledge level. First, to obtain fewer but more representative replay samples, representative images with a high number of old class foregrounds are selected by maximizing sample structure distance, providing direct foreground information at the image level. Second, the feature map output by the teacher model is used as a feature-level prompt, with a feature distillation loss guiding the student model to encode new class foreground information in less significant channels of the old feature map, reducing interference. Lastly, the teacher model’s inference output serves as a knowledge-level prompt, and the adaptive select object method is introduced to avoid foreground conflict in traditional knowledge distillation, enhancing the model’s plasticity by selectively merging foreground information. Extensive experiments on PASCAL VOC and MS COCO datasets demonstrate that this approach significantly improves model plasticity while maintaining stability. |
format | Article |
id | doaj-art-fdd2ddda95ce4f55886265a4b2a42e14 |
institution | Kabale University |
issn | 2169-3536 |
language | English |
publishDate | 2025-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj-art-fdd2ddda95ce4f55886265a4b2a42e142025-01-10T00:01:32ZengIEEEIEEE Access2169-35362025-01-01134048406610.1109/ACCESS.2024.352453910819438Multi-Level Foreground Prompt for Incremental Object DetectionJianwen Mo0https://orcid.org/0000-0002-1729-1284Ronghua Zou1https://orcid.org/0009-0008-0050-3181Hua Yuan2https://orcid.org/0009-0001-8592-9884School of Information and Communication, Guilin University of Electronic Technology, Guilin, ChinaSchool of Information and Communication, Guilin University of Electronic Technology, Guilin, ChinaSchool of Information and Communication, Guilin University of Electronic Technology, Guilin, ChinaIn the study of incremental object detection, knowledge distillation and data replay are effective methods to mitigate catastrophic forgetting. However, current research on single-stage detectors is limited, single-stage detector outputs often contain excessive negative sample information, and direct application of knowledge distillation to them is ineffective. To address this, this paper proposes a multi-level foreground prompt incremental learning algorithm for single-stage detectors like YOLO, including foreground prompts at the image level, feature map level, and knowledge level. First, to obtain fewer but more representative replay samples, representative images with a high number of old class foregrounds are selected by maximizing sample structure distance, providing direct foreground information at the image level. Second, the feature map output by the teacher model is used as a feature-level prompt, with a feature distillation loss guiding the student model to encode new class foreground information in less significant channels of the old feature map, reducing interference. Lastly, the teacher model’s inference output serves as a knowledge-level prompt, and the adaptive select object method is introduced to avoid foreground conflict in traditional knowledge distillation, enhancing the model’s plasticity by selectively merging foreground information. Extensive experiments on PASCAL VOC and MS COCO datasets demonstrate that this approach significantly improves model plasticity while maintaining stability.https://ieeexplore.ieee.org/document/10819438/Catastrophic forgettingincremental learningknowledge distillationobject detection |
spellingShingle | Jianwen Mo Ronghua Zou Hua Yuan Multi-Level Foreground Prompt for Incremental Object Detection IEEE Access Catastrophic forgetting incremental learning knowledge distillation object detection |
title | Multi-Level Foreground Prompt for Incremental Object Detection |
title_full | Multi-Level Foreground Prompt for Incremental Object Detection |
title_fullStr | Multi-Level Foreground Prompt for Incremental Object Detection |
title_full_unstemmed | Multi-Level Foreground Prompt for Incremental Object Detection |
title_short | Multi-Level Foreground Prompt for Incremental Object Detection |
title_sort | multi level foreground prompt for incremental object detection |
topic | Catastrophic forgetting incremental learning knowledge distillation object detection |
url | https://ieeexplore.ieee.org/document/10819438/ |
work_keys_str_mv | AT jianwenmo multilevelforegroundpromptforincrementalobjectdetection AT ronghuazou multilevelforegroundpromptforincrementalobjectdetection AT huayuan multilevelforegroundpromptforincrementalobjectdetection |