TY - GEN
T1 - Minimizing Inference Time
T2 - 2021 International Joint Conference on Neural Networks, IJCNN 2021
AU - Mueller, Etienne
AU - Hansjakob, Julius
AU - Auge, Daniel
AU - Knoll, Alois
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021/7/18
Y1 - 2021/7/18
N2 - Spiking neural networks offer the potential to drastically reduce energy consumption in edge devices. Unfortunately they are overshadowed by today's common analog neural networks, whose superior backpropagation-based learning algorithms frequently demonstrate superhuman performance on different tasks. The best accuracies in spiking networks are achieved by training analog networks and converting them. Still, during runtime many simulation time steps are needed until they converge. To improve the simulation time we evaluate two inference optimization algorithms and propose an additional method for error minimization. We assess them on Residual Networks of different sizes, up to ResNet101. The combination of all three is evaluated on a large scale with a RetinaNet on the COCO dataset. Our experiments show that all optimization algorithms combined can speed up the inference process by a factor of ten. Additionally, the accuracy loss between the original and the converted network is less than half a percent, which is the lowest on a complex dataset reported to date.
AB - Spiking neural networks offer the potential to drastically reduce energy consumption in edge devices. Unfortunately they are overshadowed by today's common analog neural networks, whose superior backpropagation-based learning algorithms frequently demonstrate superhuman performance on different tasks. The best accuracies in spiking networks are achieved by training analog networks and converting them. Still, during runtime many simulation time steps are needed until they converge. To improve the simulation time we evaluate two inference optimization algorithms and propose an additional method for error minimization. We assess them on Residual Networks of different sizes, up to ResNet101. The combination of all three is evaluated on a large scale with a RetinaNet on the COCO dataset. Our experiments show that all optimization algorithms combined can speed up the inference process by a factor of ten. Additionally, the accuracy loss between the original and the converted network is less than half a percent, which is the lowest on a complex dataset reported to date.
KW - conversion
KW - neuromorphic computing
KW - object detection
KW - residual networks
KW - spiking neural networks
UR - http://www.scopus.com/inward/record.url?scp=85116429694&partnerID=8YFLogxK
U2 - 10.1109/IJCNN52387.2021.9533874
DO - 10.1109/IJCNN52387.2021.9533874
M3 - Conference contribution
AN - SCOPUS:85116429694
T3 - Proceedings of the International Joint Conference on Neural Networks
BT - IJCNN 2021 - International Joint Conference on Neural Networks, Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 18 July 2021 through 22 July 2021
ER -