Abstract
Although convolutional neural networks (CNNs) are widely used in modern classifiers, they are affected by overfitting and lack robustness leading to overconfident false predictions (FPs). By preventing FPs, certain consequences (such as accidents and financial losses) can be avoided and the use of CNNs in safety- and/or mission-critical applications would be effective. In this work, we aim to improve the separability of true predictions (TPs) and FPs by enforcing the confidence determining uncertainty to be high for TPs and low for FPs. To achieve this, we must devise a suitable method. We proposed the use of Monte Carlo averaging (MCA) and thus compare it with related methods, such as baseline (single CNN), Monte Carlo dropout (MCD), ensemble, and mixture of Monte Carlo dropout (MMCD). This comparison is performed using the results of experiments conducted on four datasets with three different architectures. The results show that MCA performs as well as or even better than MMCD, which in turn performs better than baseline, ensemble, and MCD. Consequently, MCA could be used instead of MMCD for uncertainty estimation, especially because it does not require a predefined distribution and it is less expensive than MMCD.
Original language | English |
---|---|
Article number | 012004 |
Journal | Journal of Physics: Conference Series |
Volume | 2506 |
Issue number | 1 |
DOIs | |
State | Published - 2023 |
Externally published | Yes |
Event | 2022 International Joint Conference on Robotics and Artificial Intelligence, JCRAI 2022 - Virtual, Online Duration: 14 Oct 2022 → 17 Oct 2022 |
Keywords
- Convolutional neural network (CNN)
- Monte Carlo averaging (MCA)
- Monte Carlo dropout (MCD)
- confidence calibration
- ensemble
- mixture of Monte Carlo dropout (MMCD)
- separating true predictions (TPs) and false predictions (FPs)