A Measure of Confidence of Artificial Neural Network Classifiers

Konferenz: ARCS Workshop 2019 - 32nd International Conference on Architecture of Computing Systems
20.05.2019 - 21.05.2019 in Copenhagen, Denmark

Tagungsband: ARCS 2019

Seiten: 5Sprache: EnglischTyp: PDF

Persönliche VDE-Mitglieder erhalten auf diesen Artikel 10% Rabatt

Autoren:
Gschossmann, Andreas; Jobst, Simon; Mottok, Juergen; Bierl, Rudolf (Ostbayerische Technische Hochschule, Regensburg, Regensburg, Germany)

Inhalt:
Confidence in results of an Artificial Neural Networks (ANNs) is increased by preferring to reject data, that is not trustful, instead of risking a misclassification. For this purpose a model is proposed that is able to recognize data, which differs significantly from the training data, during inference. The proposed model observes all activations of the hidden layers, as well as input and output layers of an ANN in a grey-box view. To make ANNs more robust in safety critical applications, this model can be used to reject flawed data, that is suspected to decrease the accuracy of the model. If this information is logged during inference, it can be used to improve the model, by training it specifically with the missing information. An experiment on the MNIST dataset is conducted and its results are discussed.