• English
  • Deutsch
  • Log In
    Password Login
    Research Outputs
    Fundings & Projects
    Researchers
    Institutes
    Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Konferenzschrift
  4. Patch Shortcuts: Interpretable Proxy Models Efficiently Find Black-Box Vulnerabilities
 
  • Details
  • Full
Options
2021
Conference Paper
Title

Patch Shortcuts: Interpretable Proxy Models Efficiently Find Black-Box Vulnerabilities

Abstract
An important pillar for safe machine learning (ML) is the systematic mitigation of weaknesses in neural networks to afford their deployment in critical applications. An ubiquitous class of safety risks are learned shortcuts, i.e. spurious correlations a network exploits for its decisions that have no semantic connection to the actual task. Networks relying on such shortcuts bear the risk of not generalizing well to unseen inputs. Explainability methods help to uncover such network vulnerabilities. However, many of these techniques are not directly applicable if access to the network is constrained, in so-called black-box setups. These setups are prevalent when using third-party ML components. To address this constraint, we present an approach to detect learned shortcuts using an interpreta ble-by-design network as a proxy to the black-box model of interest. Leveraging the proxys guarantees on introspection we automatically extract candidates for learned shortcuts. Their transferability to the black box is validated in a systematic fashion. Concretely, as proxy model we choose a BagNet, which bases its decisions purely on local image patches. We demonstrate on the autonomous driving dataset A2D2 that extracted patch shortcuts significantly influence the black box model. By efficiently identifying such patch-based vulnerabilities, we contribute to safer ML models.
Author(s)
Rosenzweig, Julia  
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Sicking, Joachim
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Houben, Sebastian  
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Mock, Michael  
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Akila, Maram  
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Mainwork
IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2021. Proceedings  
Project(s)
KI Absicherung - Safe AI for Automated Driving
Machine-Learning Rhein-Ruhr (ML2R)
Funder
Bundesministerium für Wirtschaft und Energie BMWi (Deutschland)  
Bundesministerium für Bildung und Forschung BMBF (Deutschland)  
Conference
Conference on Computer Vision and Pattern Recognition (CVPR) 2021  
Open Access
DOI
10.1109/CVPRW53098.2021.00015
Language
English
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Contact
© 2024