Cooperative inference analysis based on DNN convolutional kernel partitioning

With the popularity of intelligent chip in the application of edge terminal devices, a large number of AI applications will be deployed on the edge of networks closer to data sources in the future.The method based on DNN partition can realize deep learning model training and deployment on resource-c...

Full description

Saved in:
Bibliographic Details
Main Authors: Jialin ZHI, Yinglei TENG, Xinyang ZHANG, Tao NIU, Mei SONG
Format: Article
Language:zho
Published: China InfoCom Media Group 2022-12-01
Series:物联网学报
Subjects:
Online Access:http://www.wlwxb.com.cn/zh/article/doi/10.11959/j.issn.2096-3750.2022.00308/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1841533698196897792
author Jialin ZHI
Yinglei TENG
Xinyang ZHANG
Tao NIU
Mei SONG
author_facet Jialin ZHI
Yinglei TENG
Xinyang ZHANG
Tao NIU
Mei SONG
author_sort Jialin ZHI
collection DOAJ
description With the popularity of intelligent chip in the application of edge terminal devices, a large number of AI applications will be deployed on the edge of networks closer to data sources in the future.The method based on DNN partition can realize deep learning model training and deployment on resource-constrained terminal devices, and solve the bottleneck problem of edge AI computing ability.Thekernel based partition method (KPM) was proposed as a new scheme on the basis of traditional workload based partition method (WPM).The quantitative analysis of inference performance was carried out from three aspects of computation FLOPS, memory consumption and communication cost respectively, and the qualitative analysis of the above two schemes was carried out from the perspective of flexibility, robustness and privacy of inference process.Finally, a software and hardware experimental platform was built, and AlexNet and VGG11 networks were implemented using PyTorch to further verify the performance advantages of the proposed scheme in terms of delay and energy consumption.It was concluded that, compared with the WPM scheme, the KPM scheme had better DNN reasoning acceleration effect in large-scale computing scenarios.And it has lower memory usage and energy consumption.
format Article
id doaj-art-203aca52d48b411bb0c5f27ded554a53
institution Kabale University
issn 2096-3750
language zho
publishDate 2022-12-01
publisher China InfoCom Media Group
record_format Article
series 物联网学报
spelling doaj-art-203aca52d48b411bb0c5f27ded554a532025-01-15T02:54:46ZzhoChina InfoCom Media Group物联网学报2096-37502022-12-016728159580497Cooperative inference analysis based on DNN convolutional kernel partitioningJialin ZHIYinglei TENGXinyang ZHANGTao NIUMei SONGWith the popularity of intelligent chip in the application of edge terminal devices, a large number of AI applications will be deployed on the edge of networks closer to data sources in the future.The method based on DNN partition can realize deep learning model training and deployment on resource-constrained terminal devices, and solve the bottleneck problem of edge AI computing ability.Thekernel based partition method (KPM) was proposed as a new scheme on the basis of traditional workload based partition method (WPM).The quantitative analysis of inference performance was carried out from three aspects of computation FLOPS, memory consumption and communication cost respectively, and the qualitative analysis of the above two schemes was carried out from the perspective of flexibility, robustness and privacy of inference process.Finally, a software and hardware experimental platform was built, and AlexNet and VGG11 networks were implemented using PyTorch to further verify the performance advantages of the proposed scheme in terms of delay and energy consumption.It was concluded that, compared with the WPM scheme, the KPM scheme had better DNN reasoning acceleration effect in large-scale computing scenarios.And it has lower memory usage and energy consumption.http://www.wlwxb.com.cn/zh/article/doi/10.11959/j.issn.2096-3750.2022.00308/edge intelligencedeep neural network partitioncooperative computationparallel partition
spellingShingle Jialin ZHI
Yinglei TENG
Xinyang ZHANG
Tao NIU
Mei SONG
Cooperative inference analysis based on DNN convolutional kernel partitioning
物联网学报
edge intelligence
deep neural network partition
cooperative computation
parallel partition
title Cooperative inference analysis based on DNN convolutional kernel partitioning
title_full Cooperative inference analysis based on DNN convolutional kernel partitioning
title_fullStr Cooperative inference analysis based on DNN convolutional kernel partitioning
title_full_unstemmed Cooperative inference analysis based on DNN convolutional kernel partitioning
title_short Cooperative inference analysis based on DNN convolutional kernel partitioning
title_sort cooperative inference analysis based on dnn convolutional kernel partitioning
topic edge intelligence
deep neural network partition
cooperative computation
parallel partition
url http://www.wlwxb.com.cn/zh/article/doi/10.11959/j.issn.2096-3750.2022.00308/
work_keys_str_mv AT jialinzhi cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning
AT yingleiteng cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning
AT xinyangzhang cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning
AT taoniu cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning
AT meisong cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning