• English
  • Deutsch
  • Log In
    Password Login
    Research Outputs
    Fundings & Projects
    Researchers
    Institutes
    Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Scopus
  4. CAN CONFORMAL PREDICTION OBTAIN MEANINGFUL SAFETY GUARANTEES FOR ML MODELS?
 
  • Details
  • Full
Options
2023
Conference Paper
Title

CAN CONFORMAL PREDICTION OBTAIN MEANINGFUL SAFETY GUARANTEES FOR ML MODELS?

Abstract
Conformal Prediction (CP) has been recently proposed as a methodology to calibrate the predictions of Machine Learning (ML) models so that they can output rigorous quantification of their uncertainties. For example, one can calibrate the predictions of an ML model into prediction sets, that guarantee to cover the ground truth class with a probability larger than a specified threshold. In this paper, we study whether CP can provide strong statistical guarantees that would be required in safety-critical applications. Our evaluation on the ImageNet demonstrates that using CP over state-of-the-art models fails to deliver the required guarantees. We corroborate our results by deriving a simple connection between the CP prediction sets and top-k accuracy.
Author(s)
Seferis, Emmanouil
Fraunhofer-Institut für Kognitive Systeme IKS  
Burton, Simon
Fraunhofer-Institut für Kognitive Systeme IKS  
Cheng, Chihhong
Fraunhofer-Institut für Kognitive Systeme IKS  
Mainwork
1st Tiny Papers Track at Iclr 2023 Tiny Papers @ Iclr 2023
Conference
1st Tiny Papers at 11th International Conference on Learning Representations, Tiny Papers @ ICLR 2023
Language
English
Fraunhofer-Institut für Kognitive Systeme IKS  
  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Contact
© 2024