Swin Transformer lightweight: an efficient strategy that combines weight sharing, distillation and pruning

Swin Transformer, as a layered visual transformer with shifted windows, has attracted extensive attention in the field of computer vision due to its exceptional modeling capabilities. However, its high computational complexity limits its applicability on devices with constrained computational resour...

Full description

Saved in:
Bibliographic Details
Main Authors: HAN Bo, ZHOU Shun, FAN Jianhua, WEI Xianglin, HU Yongyang, ZHU Yanping
Format: Article
Language:zho
Published: Beijing Xintong Media Co., Ltd 2024-09-01
Series:Dianxin kexue
Subjects:
Online Access:http://www.telecomsci.com/zh/article/doi/10.11959/j.issn.1000-0801.2024209/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1841528846270070784
author HAN Bo
ZHOU Shun
FAN Jianhua
WEI Xianglin
HU Yongyang
ZHU Yanping
author_facet HAN Bo
ZHOU Shun
FAN Jianhua
WEI Xianglin
HU Yongyang
ZHU Yanping
author_sort HAN Bo
collection DOAJ
description Swin Transformer, as a layered visual transformer with shifted windows, has attracted extensive attention in the field of computer vision due to its exceptional modeling capabilities. However, its high computational complexity limits its applicability on devices with constrained computational resources. To address this issue, a pruning compression method was proposed, integrating weight sharing and distillation. Initially, weight sharing was implemented across layers, and transformation layers were added to introduce weight transformation, thereby enhancing diversity. Subsequently, a parameter dependency mapping graph for the transformation blocks was constructed and analyzed, and a grouping matrix <italic>F</italic> was built to record the dependency relationships among all parameters and identify parameters for simultaneous pruning. Finally, distillation was then employed to restore the model’s performance. Experiments conducted on the ImageNet-Tiny-200 public dataset demonstrate that, with a reduction of 32% in model computational complexity, the proposed method only results in approximately a 3% performance degradation at minimum. It provides a solution for deploying high-performance artificial intelligence models in environments with limited computational resources.
format Article
id doaj-art-549c9e2a4bb049229af89f1090dcd120
institution Kabale University
issn 1000-0801
language zho
publishDate 2024-09-01
publisher Beijing Xintong Media Co., Ltd
record_format Article
series Dianxin kexue
spelling doaj-art-549c9e2a4bb049229af89f1090dcd1202025-01-15T03:34:01ZzhoBeijing Xintong Media Co., LtdDianxin kexue1000-08012024-09-0140667473366270Swin Transformer lightweight: an efficient strategy that combines weight sharing, distillation and pruningHAN BoZHOU ShunFAN JianhuaWEI XianglinHU YongyangZHU YanpingSwin Transformer, as a layered visual transformer with shifted windows, has attracted extensive attention in the field of computer vision due to its exceptional modeling capabilities. However, its high computational complexity limits its applicability on devices with constrained computational resources. To address this issue, a pruning compression method was proposed, integrating weight sharing and distillation. Initially, weight sharing was implemented across layers, and transformation layers were added to introduce weight transformation, thereby enhancing diversity. Subsequently, a parameter dependency mapping graph for the transformation blocks was constructed and analyzed, and a grouping matrix <italic>F</italic> was built to record the dependency relationships among all parameters and identify parameters for simultaneous pruning. Finally, distillation was then employed to restore the model’s performance. Experiments conducted on the ImageNet-Tiny-200 public dataset demonstrate that, with a reduction of 32% in model computational complexity, the proposed method only results in approximately a 3% performance degradation at minimum. It provides a solution for deploying high-performance artificial intelligence models in environments with limited computational resources.http://www.telecomsci.com/zh/article/doi/10.11959/j.issn.1000-0801.2024209/Swin Transformermodel lightweightinference accelerationpruningdistillationweight sharing
spellingShingle HAN Bo
ZHOU Shun
FAN Jianhua
WEI Xianglin
HU Yongyang
ZHU Yanping
Swin Transformer lightweight: an efficient strategy that combines weight sharing, distillation and pruning
Dianxin kexue
Swin Transformer
model lightweight
inference acceleration
pruning
distillation
weight sharing
title Swin Transformer lightweight: an efficient strategy that combines weight sharing, distillation and pruning
title_full Swin Transformer lightweight: an efficient strategy that combines weight sharing, distillation and pruning
title_fullStr Swin Transformer lightweight: an efficient strategy that combines weight sharing, distillation and pruning
title_full_unstemmed Swin Transformer lightweight: an efficient strategy that combines weight sharing, distillation and pruning
title_short Swin Transformer lightweight: an efficient strategy that combines weight sharing, distillation and pruning
title_sort swin transformer lightweight an efficient strategy that combines weight sharing distillation and pruning
topic Swin Transformer
model lightweight
inference acceleration
pruning
distillation
weight sharing
url http://www.telecomsci.com/zh/article/doi/10.11959/j.issn.1000-0801.2024209/
work_keys_str_mv AT hanbo swintransformerlightweightanefficientstrategythatcombinesweightsharingdistillationandpruning
AT zhoushun swintransformerlightweightanefficientstrategythatcombinesweightsharingdistillationandpruning
AT fanjianhua swintransformerlightweightanefficientstrategythatcombinesweightsharingdistillationandpruning
AT weixianglin swintransformerlightweightanefficientstrategythatcombinesweightsharingdistillationandpruning
AT huyongyang swintransformerlightweightanefficientstrategythatcombinesweightsharingdistillationandpruning
AT zhuyanping swintransformerlightweightanefficientstrategythatcombinesweightsharingdistillationandpruning