elib
DLR-Header
DLR-Logo -> http://www.dlr.de
DLR Portal Home | Imprint | Privacy Policy | Contact | Deutsch
Fontsize: [-] Text [+]

Regularization Strength Impact on Neural Network Ensembles

Njieutcheu Tassi, Cedrique Rovile and Boerner, Anko and Triebel, Rudolph (2023) Regularization Strength Impact on Neural Network Ensembles. In: 5th International Conference on Algorithms, Computing and Artificial Intelligence, ACAI 2022. 2022 5th International Conference on Algorithms, Computing and Artificial Intelligence, 2022-12-23 - 2022-12-25, Sanya, China. doi: 10.1145/3579654.3579661. ISBN 978-145039834-3.

[img] PDF - Only accessible within DLR
672kB

Official URL: https://dl.acm.org/doi/abs/10.1145/3579654.3579661

Abstract

In the last decade, several approaches have been proposed for regularizing deeper and wider neural networks (NNs), which is of importance in areas like image classification. It is now common practice to incorporate several regularization approaches in the training procedure of NNs. However, the impact of regularization strength on the properties of an ensemble of NNs remains unclear. For this reason, the study empirically compared the impact of NNs built based on two different regularization strengths (weak regularization (WR) and strong regularization (SR)) on the properties of an ensemble, such as the magnitude of logits, classification accuracy, calibration error, and ability to separate true predictions (TPs) and false predictions (FPs). The comparison was based on results from different experiments conducted on three different models, datasets, and architectures. Experimental results show that the increase in regularization strength 1) reduces the magnitude of logits; 2) can increase or decrease the classification accuracy depending on the dataset and/or architecture; 3) increases the calibration error; and 4) can improve or harm the separability between TPs and FPs depending on the dataset, architecture, model type and/or FP type.

Item URL in elib:https://elib.dlr.de/192934/
Document Type:Conference or Workshop Item (Speech)
Title:Regularization Strength Impact on Neural Network Ensembles
Authors:
AuthorsInstitution or Email of AuthorsAuthor's ORCID iDORCID Put Code
Njieutcheu Tassi, Cedrique RovileUNSPECIFIEDUNSPECIFIEDUNSPECIFIED
Boerner, AnkoUNSPECIFIEDhttps://orcid.org/0000-0002-7176-3588UNSPECIFIED
Triebel, RudolphUNSPECIFIEDhttps://orcid.org/0000-0002-7975-036XUNSPECIFIED
Date:March 2023
Journal or Publication Title:5th International Conference on Algorithms, Computing and Artificial Intelligence, ACAI 2022
Refereed publication:Yes
Open Access:No
Gold Open Access:No
In SCOPUS:Yes
In ISI Web of Science:No
DOI:10.1145/3579654.3579661
Series Name:ACM International Conference Proceeding Series
ISBN:978-145039834-3
Status:Published
Keywords:Ensemble, Monte Carlo Dropout (MCD), Mixture of Monte Carlo Dropout (MMCD), Regularization strength, Quality of uncertainty, Calibration error, Separating true predictions (TPs) and false predictions (FPs)
Event Title:2022 5th International Conference on Algorithms, Computing and Artificial Intelligence
Event Location:Sanya, China
Event Type:international Conference
Event Start Date:23 December 2022
Event End Date:25 December 2022
HGF - Research field:other
HGF - Program:other
HGF - Program Themes:other
DLR - Research area:Digitalisation
DLR - Program:D IAS - Innovative Autonomous Systems
DLR - Research theme (Project):D - SKIAS, R - Multisensory World Modelling (RM) [RO]
Location: Berlin-Adlershof
Institutes and Institutions:Institute of Optical Sensor Systems > Real-Time Data Processing
Institute of Robotics and Mechatronics (since 2013)
Institute of Data Science
Institute of Robotics and Mechatronics (since 2013) > Perception and Cognition
Deposited By: Njieutcheu Tassi, Cedrique Rovile
Deposited On:14 Jun 2023 12:43
Last Modified:24 Apr 2024 20:54

Repository Staff Only: item control page

Browse
Search
Help & Contact
Information
electronic library is running on EPrints 3.3.12
Website and database design: Copyright © German Aerospace Center (DLR). All rights reserved.