To Trust or Not to Trust a Regressor: Estimating and Explaining Trustworthiness of Regression Predictions

04/14/2021
by   Kim de Bie, et al.
14

In hybrid human-AI systems, users need to decide whether or not to trust an algorithmic prediction while the true error in the prediction is unknown. To accommodate such settings, we introduce RETRO-VIZ, a method for (i) estimating and (ii) explaining trustworthiness of regression predictions. It consists of RETRO, a quantitative estimate of the trustworthiness of a prediction, and VIZ, a visual explanation that helps users identify the reasons for the (lack of) trustworthiness of a prediction. We find that RETRO-scores negatively correlate with prediction error across 117 experimental settings, indicating that RETRO provides a useful measure to distinguish trustworthy predictions from untrustworthy ones. In a user study with 41 participants, we find that VIZ-explanations help users identify whether a prediction is trustworthy or not: on average, 95.1 prediction, given a pair of predictions. In addition, an average of 75.6 participants can accurately describe why a prediction seems to be (not) trustworthy. Finally, we find that the vast majority of users subjectively experience RETRO-VIZ as a useful tool to assess the trustworthiness of algorithmic predictions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/10/2023

Explaining Model Confidence Using Counterfactuals

Displaying confidence scores in human-AI interaction has been shown to h...
research
02/16/2016

"Why Should I Trust You?": Explaining the Predictions of Any Classifier

Despite widespread adoption, machine learning models remain mostly black...
research
04/26/2022

User Trust on an Explainable AI-based Medical Diagnosis Support System

Recent research has supported that system explainability improves user t...
research
07/23/2020

Are Visual Explanations Useful? A Case Study in Model-in-the-Loop Prediction

We present a randomized controlled trial for a model-in-the-loop regress...
research
01/13/2022

A Method for Controlling Extrapolation when Visualizing and Optimizing the Prediction Profiles of Statistical and Machine Learning Models

We present a novel method for controlling extrapolation in the predictio...
research
12/02/2022

Thread With Caution: Proactively Helping Users Assess and Deescalate Tension in Their Online Discussions

Incivility remains a major challenge for online discussion platforms, to...
research
12/30/2020

Human Evaluation of Spoken vs. Visual Explanations for Open-Domain QA

While research on explaining predictions of open-domain QA systems (ODQA...

Please sign up or login with your details

Forgot password? Click here to reset