TY - GEN
T1 - Regularization Strength Impact on Neural Network Ensembles
AU - Njieutcheu Tassi, Cedrique Rovile
AU - Börner, Anko
AU - Triebel, Rudolph
N1 - Publisher Copyright:
© 2022 ACM.
PY - 2022/12/23
Y1 - 2022/12/23
N2 - In the last decade, several approaches have been proposed for regularizing deeper and wider neural networks (NNs), which is of importance in areas like image classification. It is now common practice to incorporate several regularization approaches in the training procedure of NNs. However, the impact of regularization strength on the properties of an ensemble of NNs remains unclear. For this reason, the study empirically compared the impact of NNs built based on two different regularization strengths (weak regularization (WR) and strong regularization (SR)) on the properties of an ensemble, such as the magnitude of logits, classification accuracy, calibration error, and ability to separate true predictions (TPs) and false predictions (FPs). The comparison was based on results from different experiments conducted on three different models, datasets, and architectures. Experimental results show that the increase in regularization strength 1) reduces the magnitude of logits; 2) can increase or decrease the classification accuracy depending on the dataset and/or architecture; 3) increases the calibration error; and 4) can improve or harm the separability between TPs and FPs depending on the dataset, architecture, model type and/or FP type.
AB - In the last decade, several approaches have been proposed for regularizing deeper and wider neural networks (NNs), which is of importance in areas like image classification. It is now common practice to incorporate several regularization approaches in the training procedure of NNs. However, the impact of regularization strength on the properties of an ensemble of NNs remains unclear. For this reason, the study empirically compared the impact of NNs built based on two different regularization strengths (weak regularization (WR) and strong regularization (SR)) on the properties of an ensemble, such as the magnitude of logits, classification accuracy, calibration error, and ability to separate true predictions (TPs) and false predictions (FPs). The comparison was based on results from different experiments conducted on three different models, datasets, and architectures. Experimental results show that the increase in regularization strength 1) reduces the magnitude of logits; 2) can increase or decrease the classification accuracy depending on the dataset and/or architecture; 3) increases the calibration error; and 4) can improve or harm the separability between TPs and FPs depending on the dataset, architecture, model type and/or FP type.
KW - Calibration error
KW - Ensemble
KW - Mixture of Monte Carlo Dropout (MMCD)
KW - Monte Carlo Dropout (MCD)
KW - Quality of uncertainty
KW - Regularization strength
KW - Separating true predictions (TPs) and false predictions (FPs)
UR - http://www.scopus.com/inward/record.url?scp=85150375483&partnerID=8YFLogxK
U2 - 10.1145/3579654.3579661
DO - 10.1145/3579654.3579661
M3 - Conference contribution
AN - SCOPUS:85150375483
T3 - ACM International Conference Proceeding Series
BT - ACAI 2022 - Conference Proceedings
PB - Association for Computing Machinery
T2 - 5th International Conference on Algorithms, Computing and Artificial Intelligence, ACAI 2022
Y2 - 23 December 2022 through 25 December 2022
ER -