Trust Issues: Uncertainty Estimation Does Not Enable Reliable OOD Detection On Medical Tabular Data

Dennis Ulmer, Lotta Meijerink, Giovanni Cinà

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

27 Citations (Scopus)

Abstract

When deploying machine learning models in high-stakes real-world environments such as health care, it is crucial to accurately assess the uncertainty concerning a model’s prediction on abnormal inputs. However, there is a scarcity of literature analyzing this problem on medical data, especially on mixed-type tabular data such as Electronic Health Records. We close this gap by presenting a series of tests including a large variety of contemporary uncertainty estimation techniques, in order to determine whether they are able to identify out-of-distribution (OOD) patients. In contrast to previous work, we design tests on realistic and clinically relevant OOD groups, and run experiments on real-world medical data. We find that almost all techniques fail to achieve convincing results, partly disagreeing with earlier findings.

Original languageEnglish
Title of host publicationProceedings of Machine Learning Research
Pages341-354
Number of pages14
Volume136
Publication statusPublished - 2020
Externally publishedYes
Event6th Workshop on Machine Learning for Health: Advancing Healthcare for All, ML4H 2020, in conjunction with the 34th Conference on Neural Information Processing Systems, NeurIPS 2020 - Virtual, Online
Duration: 11 Dec 2020 → …

Publication series

NameProceedings of Machine Learning Research

Conference

Conference6th Workshop on Machine Learning for Health: Advancing Healthcare for All, ML4H 2020, in conjunction with the 34th Conference on Neural Information Processing Systems, NeurIPS 2020
CityVirtual, Online
Period11/12/2020 → …

Keywords

  • Electronic Health Records
  • OOD Detection
  • Uncertainty Estimation

Cite this