TY - GEN
T1 - MTL-Split
T2 - 61st ACM/IEEE Design Automation Conference, DAC 2024
AU - Capogrosso, Luigi
AU - Fraccaroli, Enrico
AU - Chakraborty, Samarjit
AU - Fummi, Franco
AU - Cristani, Marco
N1 - Publisher Copyright:
© 2024 Copyright held by the owner/author(s).
PY - 2024/11/7
Y1 - 2024/11/7
N2 - Split Computing (SC), where a Deep Neural Network (DNN) is intelligently split with a part of it deployed on an edge device and the rest on a remote server is emerging as a promising approach. It allows the power of DNNs to be leveraged for latency-sensitive applications that do not allow the entire DNN to be deployed remotely, while not having sufficient computation bandwidth available locally. In many such embedded systems scenarios, such as those in the automotive domain, computational resource constraints also necessitate Multi-Task Learning (MTL), where the same DNN is used for multiple inference tasks instead of having dedicated DNNs for each task, which would need more computing bandwidth. However, how to partition such a multi-tasking DNN to be deployed within a SC framework has not been sufficiently studied. This paper studies this problem, and MTL-Split, our novel proposed architecture, shows encouraging results on both synthetic and real-world data. The source code is available at https://github.com/intelligolabs/MTL-Split.
AB - Split Computing (SC), where a Deep Neural Network (DNN) is intelligently split with a part of it deployed on an edge device and the rest on a remote server is emerging as a promising approach. It allows the power of DNNs to be leveraged for latency-sensitive applications that do not allow the entire DNN to be deployed remotely, while not having sufficient computation bandwidth available locally. In many such embedded systems scenarios, such as those in the automotive domain, computational resource constraints also necessitate Multi-Task Learning (MTL), where the same DNN is used for multiple inference tasks instead of having dedicated DNNs for each task, which would need more computing bandwidth. However, how to partition such a multi-tasking DNN to be deployed within a SC framework has not been sufficiently studied. This paper studies this problem, and MTL-Split, our novel proposed architecture, shows encouraging results on both synthetic and real-world data. The source code is available at https://github.com/intelligolabs/MTL-Split.
KW - Deep Neural Networks
KW - Edge Devices
KW - Multi-Task Learning
KW - Split Computing
UR - http://www.scopus.com/inward/record.url?scp=85211104886&partnerID=8YFLogxK
U2 - 10.1145/3649329.3655686
DO - 10.1145/3649329.3655686
M3 - Conference contribution
AN - SCOPUS:85211104886
T3 - Proceedings - Design Automation Conference
BT - Proceedings of the 61st ACM/IEEE Design Automation Conference, DAC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 23 June 2024 through 27 June 2024
ER -