Work item:
|
F.748.20 (ex F.AI-DMPC)
|
Subject/title:
|
Technical framework for deep neural network model partition and collaborative execution
|
Status:
|
Approved on 2022-12-14 [Issued from previous study period]
|
Approval process:
|
AAP
|
Type of work item:
|
Recommendation
|
Version:
|
New
|
Equivalent number:
|
-
|
Timing:
|
-
|
Liaison:
|
-
|
Supporting members:
|
-
|
Summary:
|
Deep neural network (DNN) model inference process usually requires a large amount of computing resources and memory. Therefore, it is difficult for end devices to perform DNN models independently. It is an effective way to implement end-edge collaborative DNN execution through DNN model partition, which can reduce latency and improve resource utilization at the same time. This recommendation aims to specify the technical framework of DNN model partition and collaborative execution. First, it is necessary to predict the overall inference latency under the current system state according to different DNN partition strategies in advance. Then, choose the appropriate partition locations and collaborative execution strategy based on the equipment computation capabilities, network status and DNN model properties. Finally, implement the model collaborative execution and optimize the resource allocation in the meanwhile.
|
Comment:
|
-
|
Reference(s):
|
|
|
Historic references:
|
Contact(s):
|
|
ITU-T A.5 justification(s): |
|
|
|
First registration in the WP:
2020-06-30 16:57:17
|
Last update:
2022-11-03 11:57:02
|