Options
September 2022
Paper (Preprint, Research Paper, Review Paper, White Paper, etc.)
Titel
Transparency and Reliability Assurance Methods for Safeguarding Deep Neural Networks - A Survey
Titel Supplements
Paper presented at Workshop on Trustworthy Artificial Intelligence as a part of the ECML/PKDD 2022, September 2022, Grenoble, France
Paper published on HAL science ouverte
Abstract
In light of deep neural network applications emerging in diverse fields - e.g., industry, healthcare or finance - weaknesses and failures of these models might bare unacceptable risks. Methods are needed that enable developers to discover and mitigate such weaknesses in order to develop trustworthy Machine Learning (ML), especially in safety-critical application areas. However, it is necessary to get an insight into the rapidly developing variety of methods for correcting different deficiencies. Unlike other similar work that focuses on one particular topic, we consider three areas of action which are directly associated with the development and evaluation of ML models: transparency, uncertainty estimation and robustness. We provide an overview and comparative assessment of current approaches for building reliable and transparent models targeted at ML developers.
Project(s)
ZERTIFIZIERTE KI
Funder
Ministerium für Wirtschaft, Industrie, Klimaschutz und Energie des Landes Nordrhein-Westfalen