TY - JOUR
T1 - Monte Carlo averaging for uncertainty estimation in neural networks
AU - Tassi, Cedrique Rovile Njieutcheu
AU - Börner, Anko
AU - Triebel, Rudolph
N1 - Publisher Copyright:
© Published under licence by IOP Publishing Ltd.
PY - 2023
Y1 - 2023
N2 - Although convolutional neural networks (CNNs) are widely used in modern classifiers, they are affected by overfitting and lack robustness leading to overconfident false predictions (FPs). By preventing FPs, certain consequences (such as accidents and financial losses) can be avoided and the use of CNNs in safety- and/or mission-critical applications would be effective. In this work, we aim to improve the separability of true predictions (TPs) and FPs by enforcing the confidence determining uncertainty to be high for TPs and low for FPs. To achieve this, we must devise a suitable method. We proposed the use of Monte Carlo averaging (MCA) and thus compare it with related methods, such as baseline (single CNN), Monte Carlo dropout (MCD), ensemble, and mixture of Monte Carlo dropout (MMCD). This comparison is performed using the results of experiments conducted on four datasets with three different architectures. The results show that MCA performs as well as or even better than MMCD, which in turn performs better than baseline, ensemble, and MCD. Consequently, MCA could be used instead of MMCD for uncertainty estimation, especially because it does not require a predefined distribution and it is less expensive than MMCD.
AB - Although convolutional neural networks (CNNs) are widely used in modern classifiers, they are affected by overfitting and lack robustness leading to overconfident false predictions (FPs). By preventing FPs, certain consequences (such as accidents and financial losses) can be avoided and the use of CNNs in safety- and/or mission-critical applications would be effective. In this work, we aim to improve the separability of true predictions (TPs) and FPs by enforcing the confidence determining uncertainty to be high for TPs and low for FPs. To achieve this, we must devise a suitable method. We proposed the use of Monte Carlo averaging (MCA) and thus compare it with related methods, such as baseline (single CNN), Monte Carlo dropout (MCD), ensemble, and mixture of Monte Carlo dropout (MMCD). This comparison is performed using the results of experiments conducted on four datasets with three different architectures. The results show that MCA performs as well as or even better than MMCD, which in turn performs better than baseline, ensemble, and MCD. Consequently, MCA could be used instead of MMCD for uncertainty estimation, especially because it does not require a predefined distribution and it is less expensive than MMCD.
KW - Convolutional neural network (CNN)
KW - Monte Carlo averaging (MCA)
KW - Monte Carlo dropout (MCD)
KW - confidence calibration
KW - ensemble
KW - mixture of Monte Carlo dropout (MMCD)
KW - separating true predictions (TPs) and false predictions (FPs)
UR - http://www.scopus.com/inward/record.url?scp=85169587728&partnerID=8YFLogxK
U2 - 10.1088/1742-6596/2506/1/012004
DO - 10.1088/1742-6596/2506/1/012004
M3 - Conference article
AN - SCOPUS:85169587728
SN - 1742-6588
VL - 2506
JO - Journal of Physics: Conference Series
JF - Journal of Physics: Conference Series
IS - 1
M1 - 012004
T2 - 2022 International Joint Conference on Robotics and Artificial Intelligence, JCRAI 2022
Y2 - 14 October 2022 through 17 October 2022
ER -