Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Setting up ML Evaluation Standards to Accelerate Progress

Setting Clear Expectations for Uncertainty Estimation

Victor Bouvier · Simona Maggio · Alexandre Abraham · Dreyfus-Schmidt Schmidt


Abstract:

If Uncertainty Quantification (UQ) is crucial to achieve trustworthy ML, most UQmethods suffer from disparate and inconsistent evaluation protocols.We claim this inconsistency results from the uncleared requirements the community expects from UQ. This opinion paper offers a new perspective by specifying those requirements through five downstream tasks where we expect uncertainty scores to have substantial predictive power. On an example benchmark of 7 classification datasets, we did not observe statistical superiority of state-of-the-art intrinsic UQ methods against simple baselines. We believe that our findings question the very rationale of why we quantify uncertainty and call for a standardized protocol for UQ evaluation based on metrics proven to be relevant for the ML practitioner.

Chat is not available.