Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (2024)

Dongping Zhangdzhang@u.northwestern.edu0000-0001-9825-1411Northwestern UniversityEvanstonIllinoisUSA,Angelos Chatzimparmpasangelos.chatzimparmpas@northwestern.edu0000-0002-9079-2376Northwestern UniversityEvanstonIllinoisUSA,Negar Kamalinegar.kamali@u.northwestern.edu0000-0002-1086-6735Northwestern UniversityEvanstonIllinoisUSAandJessica Hullmanjhullman@northwestern.edu0000-0001-6826-3550Northwestern UniversityEvanstonIllinoisUSA

(2024)

Abstract.

As deep neural networks are more commonly deployed in high-stakes domains, their black-box nature makes uncertainty quantification challenging. We investigate the presentation of conformal prediction sets—a distribution-free class of methods for generating prediction sets with specified coverage—to express uncertainty in AI-advised decision-making. Through a large online experiment, we compare the utility of conformal prediction sets to displays of Top-1111 and Top-k𝑘kitalic_k predictions for AI-advised image labeling. In a pre-registered analysis, we find that the utility of prediction sets for accuracy varies with the difficulty of the task: while they result in accuracy on par with or less than Top-1111 and Top-k𝑘kitalic_k displays for easy images, prediction sets offer some advantage in assisting humans in labeling out-of-distribution (OOD) images in the setting that we studied, especially when the set size is small. Our results empirically pinpoint practical challenges of conformal prediction sets and provide implications on how to incorporate them for real-world decision-making.

conformal prediction, image labeling, semi-supervised learning, comparative user experiment

copyright: rightsretainedjournalyear: 2024doi: 10.1145/3613904.3642446conference: Proceedings of the CHI Conference on Human Factors in Computing Systems; May 11–16, 2024; Honolulu, HI, USAbooktitle: Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI ’24), May 11–16, 2024, Honolulu, HI, USAisbn: 979-8-4007-0330-0/24/05ccs: Human-centered computingHuman computer interaction (HCI)ccs: Human-centered computingEmpirical studies in HCIccs: Human-centered computingVisualization design and evaluation methodsccs: Human-centered computingEmpirical studies in visualization

1. Introduction

As machine learning models such as deep neural networks (NNs) achieve impressive predictive performance, AI-advised decision-making has become more commonplace in various situations where a human must make a decision, from high-stakes domains like medicine (e.g.,(Razzak etal., 2018)) or autonomous vehicle navigation (e.g.,(Fujiyoshi etal., 2019)) to online tasks like content moderation (e.g.,(Lai etal., 2022)) or image labeling (e.g.,(vander Wal etal., 2021)). Ideally, having access to a predictive model can help humans make more informed decisions, surpassing the capabilities of either human judgment or AI in isolation, a synergy known as complementary performance(Bansal etal., 2021; Hemmer etal., 2021). Whenever an AI system understands certain regions of the feature space better, the human counterpart can benefit by considering the model predictions. For example, in AI-advised image labeling, which we study in this paper, human labelers may benefit from viewing predictions from a model trained on previously labeled examples.

To help humans know when to trust model predictions, presenting information about the probability that the model’s prediction is correct can, in principle, be useful. For example, a decision-maker could be provided with the softmax pseudo-probability produced in the final layer of NNs for classification tasks. However, NN predictions are known to be overconfident at times(Guo etal., 2017), and heuristic measures of uncertainty are often poorly calibrated. Consequently, decision-making relying on NN predictions with the highest softmax scores (e.g., Top-1111 or Top-k𝑘kitalic_k) can fail to account for prediction error, leading to worse decisions.

Given these challenges, conformal prediction(Vovk etal., 2005; Romano etal., 2020; Angelopoulos etal., 2020) has emerged as an alternative solution that can rigorously quantify the prediction uncertainty for NNs through the use of prediction sets. Instead of presenting unreliable softmax values or other heuristics for uncertainty, the prediction set is a type of confidence interval consisting of a set of predictions with a coverage guarantee—the true class is, on average, captured within such sets with a user-specified probability (e.g., 95%percent9595\%95 %). The conformal framework is compatible with NNs because the derived sets are distribution-free—they are model-agnostic, work with finite samples, and offer non-asymptotic guarantees without distributional assumptions(Angelopoulos and Bates, 2021).

One challenge is that while conformal prediction sets express prediction uncertainty with guarantees that status quo presentations like Top-k𝑘kitalic_k predictions with softmax cannot, the conformal coverage guarantee can lead to large sets embodying high uncertainty for instances the model perceives as difficult(Babbar etal., 2022; Angelopoulos etal., 2020). Compared with Top-k𝑘kitalic_k predictions, such sets may increase cognitive load, rendering uncertainty quantification less effective(Zhou etal., 2017), thus compromising predictive performance. To understand how conformal prediction sets support human decisions aided by model predictions in an image labeling task, we contribute a large online repeated measure experiment (n=600𝑛600n=600italic_n = 600) that compares the accuracy achieved in AI-advised image labeling using conformal prediction sets with that achieved with no predictions or Top-k𝑘kitalic_k predictions. Participants in our study were tasked with labeling images from ILSVRC 2012(Deng etal., 2009) that varied in difficulty and encompassed both image stimuli that were in-distribution and out-of-distribution (OOD). This allowed us to evaluate the utility of prediction sets against Top-k𝑘kitalic_k presentation alternatives in scenarios where the model’s predictions were mostly accurate versus when the presence of “unknown unknowns” makes predictions more error-prone.

We evaluate decision-making using metrics such as accuracy and the shortest path length, which approximates the amount of deviation from the ground truth in the label space hierarchy. We use participants’ elicited willingness-to-pay at the end of the experiment to compare the perceived value of predictions against the post-hoc inferred monetary benefits derived from prediction access.

We find that for in-distribution instances, prediction sets lead to reduced labeling accuracy compared to Top-k𝑘kitalic_k predictions. However, for OOD instances where model misspecification differentially affected the calibration of Top-k𝑘kitalic_k predictions, prediction sets improve accuracy regardless of the set size. We find that participants are willing to pay roughly equivalent amounts for each type of display; if anything, they undervalue prediction sets relative to other presentations. Our results advance understanding of the strengths and limitations of prediction sets for improving AI-advised decision-making and highlight decision-makers’ tendency to over-rely on predictions, even when poorly calibrated, in an image labeling task.

2. Background

2.1. Uncertainty Quantification Using Conformal Prediction

Prior work on Human-Computer Interaction (HCI) suggests that humans can make more informed decisions when uncertainty is effectively communicated (e.g., (Joslyn and LeClerc, 2012; Kay etal., 2016; Hullman etal., 2018)), such as by presenting prediction uncertainty through static intervals(Cumming, 2014; Cumming and Finch, 2005; Manski, 2019; Taylor and Kuyatt, 1994) or animations(Hullman etal., 2015; Zhang etal., 2021). However, uncertainty quantification for NNs remains challenging(Amodei etal., 2016; Angelopoulos and Bates, 2021) due to their black-box nature and unreliability of internal heuristic confidence values, such as softmax pseudo-probabilities(Guo etal., 2017; Torralba and Efros, 2011).

One approach to quantify prediction uncertainty in classification tasks is using Bayesian NNs, where predictions depend on sampling from the posterior distributions of model parameters, carrying an inherent, mathematically grounded measure of uncertainty. However, Bayesian NNs heavily rely on distributional assumptions for their priors and can be computationally intensive to train. Conformal prediction(Vovk etal., 2005) has emerged as a more flexible and efficient solution that does not require strong model or distributional assumptions. In classification settings, conformal prediction quantifies uncertainty through prediction sets. Similar to traditional confidence intervals, the derived sets contain predicted classes and achieve a coverage guarantee that, on average, the true class is captured by the set with a user-specified probability. Formally, given a training set (x1,y1),(x2,y2),,(xn,yn)subscript𝑥1subscript𝑦1subscript𝑥2subscript𝑦2subscript𝑥𝑛subscript𝑦𝑛(x_{1},y_{1}),(x_{2},y_{2}),\dots,(x_{n},y_{n})( italic_x start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_y start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT ) , ( italic_x start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , italic_y start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT ) , … , ( italic_x start_POSTSUBSCRIPT italic_n end_POSTSUBSCRIPT , italic_y start_POSTSUBSCRIPT italic_n end_POSTSUBSCRIPT ), where xisubscript𝑥𝑖x_{i}italic_x start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT is the feature vector and yisubscript𝑦𝑖y_{i}italic_y start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT is the corresponding label, the uncertainty set function, denoted C(X)𝐶𝑋C(X)italic_C ( italic_X ), maps a feature vector xisubscript𝑥𝑖x_{i}italic_x start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT to a subset of Y={1,,K}𝑌1𝐾Y=\{1,\dots,K\}italic_Y = { 1 , … , italic_K }. This mapping ensures coverage111Recent work explores alternatives to marginal coverage, including class-conditional coverage(Gibbs etal., 2023) and relaxed versions of conditional coverage (e.g.,(Hore and Barber, 2023)). that, for any new instance x𝑥xitalic_x drawing from the same distribution as the training data, the probability P(YC(X))1α𝑃𝑌𝐶𝑋1𝛼P(Y\in C(X))\geq 1-\alphaitalic_P ( italic_Y ∈ italic_C ( italic_X ) ) ≥ 1 - italic_α over the randomness in the calibration and test points, with an error rate of α𝛼\alphaitalic_α(Angelopoulos and Bates, 2021).

If an uncertainty quantification approach is well-calibrated, we naturally expect it to reflect greater uncertainty on more difficult instances for the model (e.g., OOD instances). With prediction sets(Romano etal., 2020; Angelopoulos etal., 2020), this manifests as adaptiveness: a conformal prediction set will be larger when the task is more difficult for the model. Adaptiveness is realized through the inductive conformal method, which quantifies the prediction uncertainty of a chosen classifier using a designated calibration set(Papadopoulos etal., 2002) separate from the training data set. This set contains instances that are independently and identically distributed (i.i.d.) to those on which the model is trained.

For each calibration instance, the predicted softmax scores are adjusted using Platt scaling(Platt, 1999; Guo etal., 2017; Nixon etal., 2019) for better interpretability and further regularized to penalize unlikely classes(Angelopoulos etal., 2020). These calibrated probabilities are then ranked in descending order, reflecting the classifier’s confidence in each class. A conformity score is computed by summing these ranked probabilities up to and including the true class, capturing the model’s accumulated confidence for the instance leading up to the correct classification. From the distribution of the conformity scores of all calibration instances, a threshold is computed by taking the α𝛼\alphaitalic_α quantile after a finite sample correction for robustness. For a new and unknown instance, this calibration threshold decides the size of the prediction set by including all labels whose softmax scores exceed this calibration threshold. In our study, we opted for the Regularized Adaptive Prediction Sets (RAPS) algorithm(Angelopoulos etal., 2020) to ensure that the prediction sets of our experiment are as small as possible while achieving, on average, the desired coverage rate.

2.2. Effects of Presenting Uncertainty in AI-advised Decisions

Pairing humans with an AI system in complex decision-making scenarios can elevate performance beyond what either can achieve individually when their strengths are complementary(Bansal etal., 2021; Guo etal., 2024; Steyvers etal., 2022). In annotation tasks with large label spaces, AI models can improve accuracy and efficiency(Levy etal., 2021). Particularly, in AI-advised image labeling, various tools exist for applying automation to enhance human annotators’ speed and accuracy(Sager etal., 2021). However, performance on AI-advised tasks can be influenced by how predictions are presented—specifically, whether uncertainty is effectively communicated or not. Effective communication of uncertainty can improve the perceived trustworthiness of predictions(Zhang etal., 2020; Lai and Tan, 2019; Beller etal., 2013). Presenting a model’s accuracy is necessary to provide humans with a well-defined decision problem in many AI-advised decision settings(Hullman etal., 2024), and can encourage analytical thinking and reduce over-reliance on predictions(Zhang etal., 2020; Prabhudesai etal., 2023). However, uncertainty quantification does not guarantee improved decision-making. Cognitive biases, such as tendencies to under- or over-react to information relative to rational standards, play a major role in how uncertainty is interpreted (e.g.,(Ambuehl and Li, 2018; Tversky and Kahneman, 1971)). Additionally, uncertainty quantification can improve trust under low cognitive load, but can diminish trust when uncertainty is presented ambiguously, demanding high cognitive load(Zhou etal., 2017).

There is limited empirical evidence on how prediction sets influence trust and accuracy in AI-advised decision-making. One exception is Babbar etal. (2022), who investigated humans’ perceived utility and performance using RAPS versus Top-1111 predictions for labeling images from CIFAR-100(Krizhevsky, 2009). They found that participants using RAPS reported higher trust and perceived utility than those using Top-1111 predictions. However, their study was unlikely to have achieved sufficient power to detect differences in accuracy between the two groups due to the very small sample size (15 participants per group). They suggest that smaller sets might lead to better accuracy, but this observation was based on a comparison between sets generated by two different conformal methods. Recent work by Straitouri and Rodriguez (2023) compared labeling performance between two decision support systems: one that only allowed participants to select a label value from the prediction set and another that allowed them to freely choose one of 16 labels. Their results suggest that when the model is well-calibrated, decision-making solely based on prediction sets tends to be more reliable and accurate because of the coverage guarantee.

Motivated by overlapping research questions to Babbar etal. (2022), we conducted a much larger experiment to evaluate the utility of different uncertainty prediction displays for in-distribution and OOD stimuli, each systematically categorized by the difficulty and size of the associated prediction set. This enabled us to explore changes in performance by instance type. Contrary to the restricted agency inStraitouri and Rodriguez (2023), our study allowed participants to freely navigate a large label space, so as to allow for the possibility that even inaccurate predictions may serve as useful clues that help participants deduce the correct answer.

3. Online Experiment

3.1. Overview

We conducted a large mixed-design repeated measure experiment on Prolific. Following a pre-registered222Pre-registration: https://aspredicted.org/6gh27.pdf analysis plan, we evaluated how different displays of prediction uncertainty impacted performance on an AI-advised labeling task, including relative to performance without access to predictions. Figure1 provides an overview of our stimuli generation process, which systematically varied whether image instances were in-distribution or OOD and easy or difficult, as well as the size of conformal prediction sets.

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (1)

\Description

This figure is an overview diagram illustrating the key experimental manipulations used in the study. It visualizes the process of utilizing conformal hold-out sets to create OOD images through synthetic image corruption, categorization of these images by difficulty and set size, and the subsequent selection of task images for the labeling tasks. The diagram further depicts the interface for different treatment conditions, including labeling without predictions (e.g., the baseline condition) or with access to various prediction displays with varying uncertainty quantification, across a series of tasks involving both in-distribution and OOD images.

Each participant in our study labeled a set of 16 images sampled from the ILSVRC 2012. We assigned participants to one of four conditions representing different prediction displays: (1) no predictions (baseline), (2) Top-1111 prediction with softmax, (3) Top-10101010 predictions with softmax, or (4) the RAPS prediction set(Angelopoulos etal., 2020).

In real-world scenarios, the deployed model can encounter unexpected OOD instances where the distribution of input features differs from what was trained and calibrated. We varied whether images were in-distribution or OOD within subjects. Among the 16 stimuli, 4 were in distribution, while 12 were not, as we expected greater variation in participants’ accuracy for OOD.

In our experiment, OOD instances notably lowered the model’s Top-k𝑘kitalic_k prediction accuracy. Although the corrupted images based on the ILSVRC 2012 dataset that we used as OOD instances could compromise the conformal coverage guarantees, the adaptiveness of RAPS led to good coverage despite OOD. However, the number of labels included in the prediction sets became much larger when the instance was harder for the model. To evaluate the impact of set size on the effectiveness of prediction sets, we balanced set size within subject by categorizing images as smaller or larger based on the size of the prediction set generated for the image. This factor enabled us to examine the performance of the prediction set between size levels, as previous work suggests that a smaller set size is more preferable(Romano etal., 2020; Angelopoulos etal., 2020) and may yield higher accuracy for in-distribution images(Babbar etal., 2022). We expected smaller set sizes to be more useful in general, as they could reduce cognitive load while maintaining the conformal coverage guarantee for in-distribution images. Larger sets embodying high uncertainty could still be useful for labeling when participants were highly uncertain of what the image showed and consequently heavily relied on prediction sets. In our setting, set size could also be viewed as a nuanced measure of difficulty for in-distribution stimuli because images that required a larger set size to achieve the desired coverage were intrinsically more difficult(Angelopoulos etal., 2020).

For both in-distribution and OOD stimuli, we balanced difficulty within subject by easy and hard instances, using the median cross-entropy loss of in-distribution images as a threshold to define these levels. The former enabled us to assess whether participants were inclined to accept predictions that expressed high confidence. The latter allowed us to examine whether participants were more likely to modify or reject low-confidence predictions and instead rely on their own intuition.

Our study evaluated participants’ performance by their labeling accuracy and the shortest path length between the chosen label and the correct label in the WordNet hierarchy(Miller, 1995; Fellbaum, 1998). After they had labeled all images, we asked them to report their willingness-to-pay for the prediction display they used (i.e., a value ranging from 00 to 8888 USD, the maximum they could earn) if they were to repeat the study with a similar distribution of instances. Finally, we elicited the strategies employed to approach the tasks.

3.2. Task and Stimuli Generation

3.2.1. Overview of Stimuli Generation

To generate stimuli, we sourced 80 task images from ILSVRC 2012, evenly divided between in-distribution and OOD (i.e., 40 of each). For both in-distribution and OOD images, we further categorized each set of stimuli by difficulty (between easy and hard) and by set size (between smaller and larger). Note that because the range of observed set sizes differed depending on the difficulty of the instance, set sizes were defined relatively rather than absolutely. This categorization resulted in four distinct groups, each comprising ten images, for both in-distribution and OOD categories. Image stimuli within each group were selected to exemplify their respective categories, determined by the model’s (described below) prediction confidence and the size of the prediction set derived.

3.2.2. Predictive Model for Image Classification

Given that underperforming models are rarely deployed in real-world scenarios, we chose a classifier designed to achieve high prediction accuracy. We used a pre-trained Wide Residual Network (WRN)(Zagoruyko and Komodakis, 2016) on ILSVRC 2012. The WRN is a Residual Network (ResNet) variant(He etal., 2016), and its primary architectural distinction is the widening factor in the channel width of the convolutional layers compared to a ResNet of the same depth. Specifically, we utilized a wide_resnet101_2 model with 101101101101 layers and a widening factor of 2222, using PyTorch’s IMAGENET1K_V2 weights(Paszke etal., 2019). This model was well-calibrated for in-distribution instances, achieving a Top-1111 accuracy of 82.5%percent82.582.5\%82.5 % and a Top-10101010 accuracy of 97.9%percent97.997.9\%97.9 % over all in-distribution images, as detailed in Table6 of AppendixA. We used the WRN to create Top-1111 and Top-10101010 predictions along with their corresponding softmax pseudo-probabilities to establish prediction displays for Top-k𝑘kitalic_k conditions.

3.2.3. Calibration Set

To produce prediction sets, we randomly sampled half of the 50,000 images from the ILSVRC 2012 validation set to create a conformal calibration set, which contained 25,000 in-distribution images that were unseen by the model. This calibration set was used to perform inductive conformalization by calibrating the model based on the conformity scores derived from all images in the conformal calibration set. This process enabled the creation of prediction sets for the remaining 25,000 images in the conformal hold-out set. Our prediction sets achieved \approx95%percent9595\%95 % size-stratified conditional coverage (α=0.05𝛼0.05\alpha=0.05italic_α = 0.05), as shown in Table1. This implies that, on average, there is a 95%percent9595\%95 % probability that an image’s true label class falls within the prediction set, conditioned on all set sizes.

Sizeα=0.05𝛼0.05\alpha=0.05italic_α = 0.05
CountCoverage
0 to 113,9670.961
2 to 156,8490.946
16 to 301,5830.978
31 to 458740.975
46 to 605480.974
101 to 1,0001,1790.974
Generating OOD images

To compare the utility of different prediction displays when the classifier is poorly calibrated, we utilized OOD images, which would lower the accuracy of Top-k𝑘kitalic_k predictions and increase the size of prediction sets.To create OOD images, we systematically transformed in-distribution images from our conformal hold-out set into OOD by applying covariate shifts(Quinonero-Candela etal., 2008) through synthetic image corruption, as described by Michaelis etal. (2019). Covariate shift, in the context of supervised learning where the joint distribution over images X𝑋Xitalic_X and labels Y𝑌Yitalic_Y is P(Y|X)P(X)𝑃conditional𝑌𝑋𝑃𝑋P(Y|X)P(X)italic_P ( italic_Y | italic_X ) italic_P ( italic_X ), refers to changes in P(X)𝑃𝑋P(X)italic_P ( italic_X ), the feature probability, without affecting the classifier P(Y|X)𝑃conditional𝑌𝑋P(Y|X)italic_P ( italic_Y | italic_X ).

After evaluating 15 types of corruption, we opted to use five corruptions that reduced WRN’s Top-1111 prediction accuracy the most. These corruptions were: (1) defocus blur, (2) snow, (3) frost, (4) zoom blur, and (5) glass blur, which are bolded in Table6 of AppendixA. Given that the human visual system can be resilient to certain minor changes(Michaelis etal., 2019; Hendrycks and Dietterich, 2019), we applied these corruptions to maximum severity for all images in the conformal hold-out set, as defined byMichaelis etal. (2019).

Table6 of Appendix A indicates that our prediction sets achieve a 96%percent9696\%96 % coverage rate for in-distribution images. However, the coverage guarantee for the OOD samples was compromised, as the images used during calibration and testing are no longer exchangeable (i.i.d.), according to Angelopoulos etal. (2020).

Classifying stimuli by difficulty

To compare the effects of different prediction displays on images of varying difficulty, we categorized images by difficulty using the cross-entropy loss, which calculates how well the predicted probability distribution aligns with the ground truth distribution. Formally in Equation1, let x𝑥xitalic_x be an input image; p(x)𝑝𝑥p(x)italic_p ( italic_x ) denotes the true probability distribution of the image’s label in one-hot encoding, while q(x)𝑞𝑥q(x)italic_q ( italic_x ) is our classifier’s predicted distribution.

(1)H(p,q)=p(x)log(q(x))𝐻𝑝𝑞𝑝𝑥𝑞𝑥H(p,q)=-\sum{p(x)\cdot\log(q(x))}italic_H ( italic_p , italic_q ) = - ∑ italic_p ( italic_x ) ⋅ roman_log ( italic_q ( italic_x ) )

Figure9 of Appendix A displays the distribution of cross-entropy loss of all in-distribution image stimuli, with the median presented as an orange dotted line. Due to WRN’s high predictive power, this distribution was highly right-skewed, which created a problem: if we used this median threshold to categorize difficulty, we would include many images in the easy group that the model was both confident and accurate in predictions, so the derived sets embodied little uncertainty, synonymous with Top-1111 prediction. We thus computed a new median threshold (0.720.720.720.72) after excluding images with a set size of one such that the difficulty threshold, shown as a red solid line, was defined based on images with non-deterministic sets. Images with cross-entropy loss falling below or on this threshold were classified as easy, while those above were deemed hard.

Because the cross-entropy loss derived from the predictive distribution of OOD images can be misleading as a signal of task difficulty due to model misspecification(Guo etal., 2017; Torralba and Efros, 2011), we used the red median threshold from the in-distribution data to categorize the difficulty of OOD images.

Hold-out Setw/ Size = 1w/o Size = 1
EasyHardEasyHard
None115519
defocus blur259860
snow141743
frost246850
zoom blur2441047
glass blur263965
Classifying stimuli by set size

Images within each difficulty group of the conformal hold-out set induce varied prediction set sizes. For instance, while images classified as easy typically had a smaller set size, some easy instances required a larger set size to achieve the desired coverage. To explore how set size influences the efficacy of prediction sets, we categorized the stimuli within each difficulty group by their set size. After grouping images in each conformal hold-out set by easy and hard, we further distinguished them based on relative set size, as determined by RAPS, into smaller or larger categories, based on the median of the group. However, as shown in Table2, the predictive power of WRN consistently demonstrated high confidence in its predictions for many easy images with a set size of one, which significantly lowered the median size threshold. To introduce more variance in set size within the easy categories, we calculated the median threshold after excluding images with a set size of one (i.e., these were directly assigned to the smaller group). Images with a set size smaller than or equal to this adjusted size median were classified as smaller; if the opposite is true, they were classified as larger. The specific thresholds used for each difficulty category within the conformal hold-out set are outlined in Table2, with the bold figures representing the applied thresholds.

Sampling task images based on defined categories

Upon categorizing all images in each conformal hold-out set (i.e., one in-distribution and five OOD, totaling 6 ×\times× 25,000 images) by difficulty (easy and hard) and size (smaller and larger), our objective was to select a set of task images from which to sample image assignments for participants, so as to avoid results that overfit to a very small set of stimuli. Specifically, we targeted 80 task images from the 1,200 candidate images. Our goals were to ensure that: (1) the candidate images are representative of each group defined by in-distribution or OOD, difficulty, and size (6×2×26226\times 2\times 26 × 2 × 2), and (2) task images avoid the known data anomalies in ILSVRC 2012(Ekambaram etal., 2017; Northcutt etal., 2021).

ShiftDifficultySizeCountCoverageAvg. Set Size
Top-1Top-10RAPS
ineasysmaller101112.6
larger1011119.5
hardsmaller100.410.98.0
larger100.50.90.951.5
outeasysmaller101115
larger1011128.1
hardsmaller1000.60.930.1
larger100.10.20.990.8

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (2)

\Description

This figure displays a grid showcasing example stimuli participants viewed for labeling tasks. Rows differentiate in-distribution and OOD stimuli, while columns further categorize stimuli examples by difficulty and set size.

Sampling images from groups, such as in-distribution, easy (difficulty), and smaller (set size), was challenging due to the prevalence of images with a set size of one for which prediction sets become nearly identical to the Top-1111 prediction. To avoid oversampling sets of size one, (1) we temporarily set aside images with a set size of one, (2) we took images with sets larger than one whose set size falls near the median size in that group (i.e., 45454545th and 55555555th percentiles), and (3) we performed a post-hoc sampling correction to include images with a set size of one in proportion to their original presence in each group, as shown in Table8 of AppendixA.

Images from ILSVRC 2012 are known to be noisy, with numerous data anomalies. To create the set of 80 task images, we sampled 50 candidate task images from each representative subset defined by in-distribution or OOD, difficulty, and size (6×2×2×50622506\times 2\times 2\times 506 × 2 × 2 × 50) without replacement. We manually inspected the task images, starting from images with higher cross-entropy loss (hard), and omitted those with (1) an obviously wrong ground truth label, (2) a text overlay with the correct label, (3) no apparent focal object or many surrounding objects, (4) unusual dimensions, and (5) small multiples. Based on these rules, we selected 10 images from each of the 4 groups from the combination of [easy,hard]×[smaller,larger]easyhardsmallerlarger[\texttt{easy},\texttt{hard}]\times[\texttt{smaller},\texttt{larger}][ easy , hard ] × [ smaller , larger ] that were in-distribution, and 2 images from each of the 4 groups with corruption. This process generated 80 task images in total, balanced by in-distribution or OOD, difficulty, and set size.

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (3)

\Description

This figure contains screenshots of the interface used by participants to complete the study, organized by conditions: baseline, Top-1, Top-10, and RAPS. Each screenshot of the interface illustrates the layout, adapted to accommodate different prediction displays with varying uncertainty quantification.

Table3 presents the Top-k𝑘kitalic_k accuracy, coverage, and average set size, grouped by our experimental factors, with example stimuli shown in Figure2. Our stimuli exhibited the desired behavior, with the average prediction accuracy and set size varying in response to the manipulated difficulty and set size. While the stimuli we selected to achieve balance over our experimental manipulations resulted in close conformal coverage for OOD instances to the in-distribution case, anecdotally we found that OOD coverage tended to remain high (\approx80%percent8080\%80 %) over varying stimuli sets under our data generating model.

3.2.4. Reward and Bonus

Participants earned a base compensation of $4currency-dollar4\$4$ 4 for completing the study and received a $0.25currency-dollar0.25\$0.25$ 0.25 bonus for each correct label. With 16 trials, the maximum bonus was $4currency-dollar4\$4$ 4, bringing the total potential earnings to $8currency-dollar8\$8$ 8. We determined the task reward to achieve the local minimum wage based on the assumption that most participants would take longer than the average time we observed in a small pilot of the experiment.

3.3. Experimental Interface

Our task interface, screenshots of which are presented in Figure3, featured a two-column design: task images were on the left, and predictions were on the right, with a response field directly below the task image. In the baseline condition without predictions, the task image and the response field were centered. Participants could view their real-time earnings in the top-right corner of the interface.

To ensure that participants understood the meaning of each label, we selected label-representative images from the conformal calibration set. To find suitable label-representatives, we manually examined all images by labels and prioritized selection for those with small cross-entropy loss (i.e., easy), applying the same rules used when selecting task images. The label-representative images we used are included in the Supplemental Material.

3.3.1. Presenting Predictions

For the three treatment conditions that featured predictions, the Top-1111 condition presented the classifier’s Top-1 prediction and its softmax score on top of a representative image of the predicted label. Top-10101010 predictions with softmax scores were ordered in a table, while the prediction set was presented in an n𝑛nitalic_n-by-3333 matrix sorted column-wise by the classifier’s prediction confidence. Participants could hover their mouse cursor over any predicted label to view a label-representative image.

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (4)

\Description

This figure contains screenshots to demonstrate the search features of the interface provided to participants to assist with their labeling tasks. When performing a dropdown search, a dropdown menu will appear below the input field, which shows potential labels that match what the participants typed. Keyword search enables participants to search for labels that match their typed keywords by clicking a “Search” button. Bottom-up search allows participants to explore the label space by clicking on a predicted label. Additionally, participants can hover over any leaf node in the hierarchy network to access label-representative images.

3.3.2. Label Space Search

To label task images, participants need to be aware of the label space. Because ILSVRC 2012 has a large label space of 1,000, cumbersome to present in a dropdown or checkbox format, we organized labels using a version of the WordNet hierarchy, which we truncated slightly to avoid label overlap. For example, “automobile” and “truck” are both parent nodes for “minivan” and since “truck” is also a type of “automobile”, we retained “minivan” in the “automobile” category. We include this tree in the Supplemental Material.

If participants disagreed with the predictions or when no prediction was available (baseline), our interface enabled them to perform three types of search. As illustrated in Figure4, they were (A) dropdown search, (B) keyword search, and (C) bottom-up search. In the dropdown search, a dropdown menu would appear as participants began typing into the response field, suggesting leaf labels that matched their input. Alternatively, in the keyword search, participants could type and then click a search button to search the hierarchy for specific keywords. Doing so would open an overlay window showing the label-space hierarchy network, where leaf nodes (i.e., denoting valid labels) would appear in a different color from the parent nodes. Clicking on a leaf node would populate the label in the response field. To ease navigation, the hierarchy network was collapsed by default except for the parent nodes (i.e., categories) and leaf nodes (i.e., labels) containing the search keywords with edges connecting them highlighted in orange. All other parent nodes could also be clicked to expand.

Participants could also perform a bottom-up search starting from the prediction display. Clicking on a predicted label would open the search overlay. Instead of showing the entire hierarchy network, the overlay presented the path from the root node to the leaf node corresponding to the predicted label that the participant clicked on. Participants could explore adjacent labels under the same parent category or explore other parent nodes on the path toward the root.

When exploring the label space hierarchy, participants could hover over any leaf node to display its label-representative image. We illustrate these search actions through video demonstrations included in the Supplemental Material.

3.4. Experimental Procedure

Participants were directed to our study interface from Prolific. On the welcome page, they received a brief description of the study’s purpose and the estimated time required for completion. Participants were instructed to complete the study in a single session using Google Chrome on a large-screen device. If participants agreed to the terms by clicking the “Start” button, they would be randomly assigned to a prediction display condition with task images sampled from each corresponding group, shuffled in a randomized order.

Participants then reviewed two instruction pages. The first introduced the interface layout and detailed the bonus mechanism. The second contained several short videos demonstrating how to explore predictions (except in the baseline condition) and navigate the label space. Participants had to watch all provided instructional videos before proceeding to an example page where they could practice using the interface. Then, they completed 16 rounds of tasks. Upon finishing the 16 trials, participants were asked to state their willingness-to-pay to have access to the same style of prediction display they used in the experiment on a scale from 00 to 8888 USD (i.e., the range of possible earnings from the experiment) if they were invited to undertake another 16 rounds of tasks with new images generated similarly to those they had experienced, for further reward and potential bonus. On the same page, participants could optionally describe their strategy to identify the correct label for each image, with or without the aid of predictions.

3.5. Participants

We recruited 600 participants from Prolific, divided equally among the four prediction displays, ensuring a gender-balanced sample. We set additional screening criteria so that all of our participants were (1) based in the USA, (2) had no vision impairment, (3) had no colorblindness, (4) spoke English as the first language, and (5) aged between 18 and 65. This led to 150 participants in each condition representing different prediction displays. We determined the target sample size using a non-parametric simulation of the experiment bootstrapped from a pilot study of 30 participants. By grouping trial-level responses according to all manipulated factors and bootstrapping these responses with replacement, we simulated trial-level outcomes. We identified the target sample size (600) that resulted in 95%percent9595\%95 % confidence intervals on accuracy with widths less than 10%percent1010\%10 %.

3.6. Analysis Method

We pre-registered an analysis plan focusing on three response variables: (1) accuracy, (2) shortest path length, and (3) willingness-to-pay, and followed our pre-registered plan in all analyses reported below unless otherwise specified. We define accuracy (ACC) as binary, indicating whether a participant selected the correct label for the task. The shortest path (SP) length is a continuous variable representing the number of hops from the leaf node a participant selects to the leaf node of the ground truth label. SP length is a measure widely used to quantify semantic similarity(Resnik, 1995; Budanitsky and Hirst, 2006) between categories in a taxonomy. Although highly correlated with accuracy, this measure provides further information on how “incorrectness” varied by our experimental manipulations. Willingness-to-pay (WTP) ranges from $0currency-dollar0\$0$ 0 to $8currency-dollar8\$8$ 8, measuring how much they will pay to access the display.

We pre-registered Bayesian Linear Mixed-effect models for accuracy and SP length. Our model specifications are presented in Model 1 and Model 2. In this context, trial is numeric, indicating the trial number, whereas condition (4 levels: baseline, Top-1111, Top-10101010, and prediction set), shift (2 levels: in-distribution and OOD), difficulty (2 levels: easy and hard), and size (2 levels: smaller and larger) are factors corresponding to our experimental manipulations. The term ID represents each participant’s unique Prolific ID.

1acc similar-to\sim Bernoulli(p𝑝pitalic_p)

logit(p𝑝pitalic_p)=condition \ast shift \ast difficulty \ast size \ast trial + (trial||||ID)

For the accuracy model stated in Model 1, we use a Bernoulli distribution for the likelihood shown in Line1, parameterized by p𝑝pitalic_p, which denotes the probability of a correct label. Line 1 presents the hierarchical logistic regression model. In this specification, we model the logit of p𝑝pitalic_p through interactions among all fixed effects and incorporate random intercepts and slopes for trial, grouped by participants’ unique ID. This approach allows us to account for participants’ baseline performance differences and variations in rates of change across conditions and trials.

1sp similar-to\sim ZeroInflatedNegBinomial(μ𝜇\muitalic_μ, θ𝜃\thetaitalic_θ, ψ𝜓\psiitalic_ψ)

2log(μ𝜇\muitalic_μ)=condition \ast shift \ast difficulty \ast size \ast trial + (trialID)

logit(ψ𝜓\psiitalic_ψ)=condition \ast shift \ast difficulty \ast size \ast trial

For the SP length model stated in Model 2, we employ a Zero-inflated Negative Binomial (ZINB) model as the likelihood, parameterized by μ𝜇\muitalic_μ (expected SP length), θ𝜃\thetaitalic_θ (negative binomial dispersion), and ψ𝜓\psiitalic_ψ (zero-inflation probability), as shown in Line 2. The expected path length between the participants’ responses and the truth label is modeled by the negative binomial component in Line2. We apply a log-link function to μ𝜇\muitalic_μ, modeled by all fixed effects along with their interactions with random intercepts and slopes for trials grouped by participant ID. We do not explicitly model the dispersion parameter for the NB component. However, we placed an exponential prior with a rate of 0.10.10.10.1 to moderate the degree of dispersion in our model. Given that our observed data had a maximum SP length of 27, this prior acts as a regularizing component, ensuring our model captures the dispersion of the observed data without predicting implausibly large SP lengths. Line2 shows the zero-inflation component that models the proportion of zero path length (i.e., correct answer) not captured by the NB components. We apply a logit link function to ψ𝜓\psiitalic_ψ and model it by all fixed effects along with their interactions.

We deviated from our pre-registered model specifications only in setting the priors. We initially aimed to set weakly informative priors, but several were too narrow to allow model convergence. We therefore widened the priors to be even less informative. Full details are available in the Supplemental Material.

We followed a standard Bayesian workflow(Gelman etal., 2020) to check if both models fit. Detailed model diagnostics are included in the Supplemental Material. We use the models to generate predictions of the respective response variable based on 1,000 draws from the posterior distribution of the models’ fixed effects for each unique combination of our experimental manipulations while holding the trial effect constant at the mean. Our results present the expected predictions as point estimates, with uncertainty expressed through the 95%percent9595\%95 % highest posterior density interval (HDI).

We pre-registered an analysis plan to assess participants’ reported WTP, relative to the aggregate observed value of each prediction display. Unlike traditional Likert-style questions, WTP allows us to define an objective benchmark post-hoc for comparison to participants’ responses.

Specifically, we first calculate the expected bonus participants gained in each condition by multiplying the expected accuracy rates by the maximum earnable bonus of $4currency-dollar4\$4$ 4. We then calculate the expected bonus difference with access to each prediction display relative to the baseline by subtracting the expected bonus of the baseline from the expected bonus of each treatment condition. We finally evaluate the discrepancy between this expected bonus difference and the corresponding WTPs reported by participants who used each prediction display by constructing a “willingness-to-overpay” metric. This measure reflects the difference between the additional bonus one is expected to earn from having access to a prediction display over the baseline and the actual difference in the average bonus participants received from accessing a prediction display relative to the baseline. We use bootstrapping to derive 95% confidence intervals with the median as a point estimate.

3.7. Qualitative Analysis of Strategies

We performed a qualitative analysis and developed a bottom-up open coding scheme using the strategies provided by the participants. Two coders worked in collaboration, each coding half of the strategies. After coding, any ambiguous strategies were discussed until mutual agreements were reached. Among 585 valid strategies, we excluded 41 uninformative ones (Code: 1) with responses such as “I have no idea, trying my hardest”. For the remaining strategies, we identified 17 ways in which the prediction displays and the search features were used, as illustrated in Table4.

CodeCategory
2–6How do participants use different prediction displays?
7–10Do participants find predictions to be trustworthy?
11–15How do participants use the search feature to find a preferred label?
16–18How do participants narrow down their choices and identify a preferred label?

We provide detailed code descriptions in Table9 of AppendixA, and generate a spreadsheet of codes and quotes to summarize the diversity and nuances in participants’ strategies, which are included in the Supplemental Material.

4. Results

4.1. Data Preliminaries

We received 599 valid responses after excluding one participant from the baseline condition according to our pre-registered exclusion rule. As demonstrated in Figure5 left, participants spent, on average, \approx20202020 minutes completing the 16 image labeling tasks. Participants in the Top-10 and RAPS conditions took slightly longer to complete tasks than those in the baseline and Top-1 conditions.

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (5)

\Description

This figure presents the distributions of participants’ study completion time and the amount of bonus earned by treatment conditions. Each distribution has a solid vertical line showing the average and two dotted vertical lines showing the interquartile range.

From the distribution of earned bonuses shown in Figure5 right, there is a clear difference indicating that access to the prediction displays helped participants achieve more correct answers and hence a higher bonus reward than those without access. On average, baseline participants provided seven correct labels, which is lower than those who used prediction displays (10 out of 16 correct).

4.2. Accuracy

4.2.1. Overview

In Figure6, we present the expected predictions by our accuracy model (Model 1) for each type of image stimuli, with uncertainty quantified as 95%percent9595\%95 % HDIs. At a high level, we find that RAPS does not improve participants’ labeling accuracy for easy tasks but is more useful for hard tasks, especially for OOD images. While a smaller RAPS can be more useful for in-distribution images, size does not appear to affect accuracy much when images are OOD. Because smaller and larger set sizes depend on whether instances are in-distribution or OOD in our data-generating model (Section3.2), we provide the average set size for size groups to contextualize the results below.

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (6)

\Description

This figure displays a grid presenting the median expected labeling accuracy predicted by our accuracy model with uncertainty expressed as 95% HDIs. Rows differentiate in-distribution and OOD results, while columns separate results by difficulty and set size. Solid horizontal lines for each stimuli group illustrate the prediction accuracy achieved by that display (i.e., the probability that the displayed predictions included the true label) using our base classifier (i.e., WRN).

4.2.2. In-distribution

For both easy and hard images that are in-distribution, a larger set size may decrease participants’ labeling accuracy

This effect is most prominent when images are classified as hard in Figure6, (C) and (D). When using RAPS to label hard images, participants’ accuracy is higher when the set size is smaller (74.2%percent74.274.2\%74.2 %; HDI: [66.5%,81.3%]percent66.5percent81.3[66.5\%,81.3\%][ 66.5 % , 81.3 % ]; average size: 2.62.62.62.6), and lower when set size is larger (52.8%percent52.852.8\%52.8 %; HDI: [43.5%,60.9%]percent43.5percent60.9[43.5\%,60.9\%][ 43.5 % , 60.9 % ]; average size: 19.519.519.519.5). We speculate that the reason is that a larger set size increases participants’ cognitive load as they need to traverse through the set and evaluate all labels. Figure6 (C) shows that expected accuracy of RAPS participants is roughly 8888% higher than that of the Top-10 condition (64.5%percent64.564.5\%64.5 %; HDI: [56.4%,71.7%]percent56.4percent71.7[56.4\%,71.7\%][ 56.4 % , 71.7 % ]) which is roughly 10101010% higher than the Top-1 condition (55.6%percent55.655.6\%55.6 %; HDI: [47.2%,63.7%]percent47.2percent63.7[47.2\%,63.7\%][ 47.2 % , 63.7 % ]), though HDIs overlap.

When the true label is in the prediction display, participants do not always choose it.

This is particularly evident for in-distribution images categorized as easy (cf.Figure6, A and B). Here, the participants’ labeling accuracy tends to be lower than the prediction display accuracy (i.e., the percentage of the time the prediction display included the true label), depicted as solid horizontal lines. However, when labeling hard images, participants in the Top-1 condition can improve upon the model’s prediction. As suggested by our open codes discussed in Section4.5, we find that when the Top-1 prediction was obviously inaccurate, participants were more likely to rely on their own judgment and used the search tool to identify the correct answer. The presence of a greater number of labels in the Top-10 and RAPS conditions may have complicated the labeling process due to an increased cognitive load. This is exemplified by the observed trends in these conditions: among incorrect responses, participants in the RAPS condition missed the correct label from the prediction set 60%percent6060\%60 % of the time, whereas those in the Top-10 condition did so 80%percent8080\%80 % of the time.

In summary, when task images are in-distribution, (1) labeling accuracy for participants in the Top-1 or Top-10 conditions is higher when the images are easy (cf. Figure6, A and B); (2) the accuracy of RAPS participants is negatively correlated with set size: a larger set size decreases accuracy, particularly for hard images (cf. Figure6, C and D); and (3) Top-1 participants are more likely to rely on their own judgment and use the search tool to identify the correct answer when predictions are less likely to be correct. In contrast, the pattern of incorrect responses observed in the Top-10 and RAPS conditions suggest that the presence of more predictions makes it harder for participants to discern between inaccurate and accurate predictions, even when the correct label is present.

4.2.3. Out-of-distribution

When OOD images are easy, Top-k𝑘kbold_italic_k predictions yield higher accuracy than RAPS

Due to the similarity in HDIs across conditions for easy OOD images (cf. Figure6, E and F), we report expected accuracy marginalized over set sizes. We find Top-1 participants tend to achieve slightly higher accuracy in expectation (89.6%percent89.689.6\%89.6 %; HPI: [85.5%,93.4%]percent85.5percent93.4[85.5\%,93.4\%][ 85.5 % , 93.4 % ]) than those using Top-10 (85.7%percent85.785.7\%85.7 %; HDP: [81%,89.6%]percent81percent89.6[81\%,89.6\%][ 81 % , 89.6 % ]). Meanwhile, the RAPS condition yields the lowest expected accuracy of 79.5%percent79.579.5\%79.5 % (HPI: [75.1%,83.6%]percent75.1percent83.6[75.1\%,83.6\%][ 75.1 % , 83.6 % ]), despite an average set size of 5555 (as detailed in Table3) that is smaller than the size of Top-10, with identical chance that the prediction display contained the true label (i.e., colored horizontal lines in Figure6, E and F). This divergence from our in-distribution finding, where smaller set sizes typically correlate with higher accuracy when predictions are reliable, suggests that additional factors may be at play in the OOD scenario. It is possible that the reduction in cognitive load afforded by having access to softmax scores in the Top-10 condition aided participants in finding the correct label. This observation partially reinforces our finding from the easy in-distribution results: when the model is well-calibrated (cf. Figure6, E and F), viewing fewer predictions and having access to prediction confidence (i.e., as in our Top-k𝑘kitalic_k conditions) can improve accuracy.

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (7)

\Description

This figure displays a grid presenting the median expected shortest path length predicted by our shortest path model with uncertainty expressed as 95% HDIs. Rows differentiate in-distribution and OOD results, while columns separate results by difficulty and set size.

When OOD images are hard, RAPS yield the highest labeling accuracy regardless of the set size

For hard OOD images, RAPS set sizes grow substantially larger, with smaller sets averaging 30303030 instances and larger sets averaging 91919191 instances. When hard OOD images had smaller set sizes (i.e., Figure6, G), RAPS participants outperform those in Top-k𝑘kitalic_k conditions with a relatively high accuracy of 45.1%percent45.145.1\%45.1 % (HPI: [41%,51.6%]percent41percent51.6[41\%,51.6\%][ 41 % , 51.6 % ]), followed by the Top-10 condition (36.3%percent36.336.3\%36.3 %; HPI: [31.4%,41.3%]percent31.4percent41.3[31.4\%,41.3\%][ 31.4 % , 41.3 % ]) and Top-1 condition (16.9%percent16.916.9\%16.9 %; HPI: [13.5%,20.7%]percent13.5percent20.7[13.5\%,20.7\%][ 13.5 % , 20.7 % ]). When the set size is larger (i.e., Figure6, H), we identify a consistent pattern that RAPS participants (51.1%percent51.151.1\%51.1 %, HPI: [45.9%,56.5%]percent45.9percent56.5[45.9\%,56.5\%][ 45.9 % , 56.5 % ]) outperform those in the Top-k𝑘kitalic_k conditions. Nevertheless, the accuracy of RAPS participants is still lower than the average coverage rate (cf. Figure6, G and F, with green solid horizontal lines), implying that even when the sets include the true label, it is still challenging for participants to identify it.

When OOD images are hard, participants relying on their own judgment may achieve better accuracy than those relying on poorly-calibrated Top-k𝑘kbold_italic_k predictions

For hard OOD images with smaller average set size (i.e., Figure6, G), we find that participants in the baseline condition who did not have access to any predictions have a higher accuracy of 23.6%percent23.623.6\%23.6 % (HPI: [19.3%,27.6%]percent19.3percent27.6[19.3\%,27.6\%][ 19.3 % , 27.6 % ]) than those in the Top-1 condition (16.9%percent16.916.9\%16.9 %; HPI: [13.5%,20.7%]percent13.5percent20.7[13.5\%,20.7\%][ 13.5 % , 20.7 % ]). For hard OOD images with a larger average set size (i.e., Figure6, H), the accuracy of using RAPS (51.1%percent51.151.1\%51.1 %, HPI: [45.9%,56.5%]percent45.9percent56.5[45.9\%,56.5\%][ 45.9 % , 56.5 % ]) versus the baseline condition (52.2%percent52.252.2\%52.2 %; HPI: [47%,57.7%]percent47percent57.7[47\%,57.7\%][ 47 % , 57.7 % ]) is similar, and participants who used Top-1 prediction are more accurate (40.9%percent40.940.9\%40.9 %; HPI: [36.1%,46%]percent36.1percent46[36.1\%,46\%][ 36.1 % , 46 % ]) than those who used Top-10 predictions (30.1%percent30.130.1\%30.1 %; HPI: [25.3%,34.4%]percent25.3percent34.4[25.3\%,34.4\%][ 25.3 % , 34.4 % ]).Our results suggest that Top-k𝑘kitalic_k predictions from a poorly calibrated model (cf. Figure6, G and H) can negatively influence human judgment relative to no model assistance. Although RAPS participants have relatively higher accuracy than those in the Top-k𝑘kitalic_k conditions, their accuracy is almost identical to those in baseline, suggesting it might be beneficial to withhold predictions for hard OOD stimuli that require a larger set size to achieve the desired coverage, at least in an OOD setting like ours, where the perturbations decreased both human and AI performance.

In summary, (1) when OOD images are easy and the predictions and softmax scores well-calibrated (cf. Figure6, E and F), Top-k𝑘kitalic_k predictions tend to result in higher labeling accuracy; (2) when OOD images are hard and the model predictions more error-prone (cf. Figure6, G and H), RAPS participants can outperform their Top-k𝑘kitalic_k counterparts. Unlike in the in-distribution case, increasing the size of the prediction set does not degrade accuracy; and (3) when OOD images are hard and the model predictions are more error-prone, participants who rely on their own judgment and the label space search tool may be more likely to identify the correct answer. The reasons behind the performance advantage of the participants in the baseline condition when labeling hard OOD images with larger set sizes remain ambiguous. It is possible that the accuracy decrease for participants in Top-10 condition when labeling hard OOD images that have larger set size is due to overreliance on inaccurate predictions, given that the average coverage for Top-10 predictions decreases from 60%percent6060\%60 % (smaller) to 20%percent2020\%20 % (larger), as shown in Table3.

4.3. Shortest Path Length

The predictions of our SP length model (Model 2) are highly correlated with those of the accuracy model when comparing the results in Figure6 with those in Figure7. We briefly highlight the main takeaways that complement our accuracy results, with more detailed descriptions in AppendixB.2.

When OOD images are hard, a larger prediction set leads to submitted labels that are slightly closer to the correct label

Recall that in Section4.2.3, we find when OOD images are classified as hard, RAPS participants can outperform their Top-k𝑘kitalic_k counterparts, but there is no difference in accuracy by set size (cf. Figure6, G and H). However, from the SP length, we find that if the task image has a larger set size, RAPS participants’ chosen label is slightly closer to the correct answer in the label space hierarchy. As demonstrated inFigure7 (G) and (H), the SP length for hard OOD images that have a larger set size is 3.43.43.43.4 (HPI: [3.1,3.7]3.13.7[3.1,3.7][ 3.1 , 3.7 ], average size: 90.890.890.890.8) and that of a smaller set size is 4.2 (HPI: [3.8,4.7]3.84.7[3.8,4.7][ 3.8 , 4.7 ]; average size: 30.130.130.130.1). It is difficult to say why this might be the case; it is possible, as supported by our open codes in Section4.5, that when OOD stimuli are especially challenging to recognize, participants spend more time scrutinizing labels in the prediction set for clues, and that even when the probability that the prediction set contains the true label is the same, viewing more predictions helps cue their ability to identify the true label. The effect is quite small, whatever the cause.

When predictions are generated by a well-calibrated model, having access to predictions can guide participants closer to the correct answer

This result is most prominent for hard in-distribution images with smaller set size. From Figure6 (C), we see the baseline and Top-1 prediction tend to yield similar accuracy, as HDIs closely overlap. However, in Figure7 (C), we find that participants who used Top-1 prediction can provide responses that are closer to the correct answer than those in the baseline.

In summary, we find (1) for hard OOD stimuli in our setting, although the accuracy of RAPS participants is similar, a larger set may guide responses closer to the ground truth (cf. Figure7, G and H); and (2) when the model is well-calibrated, inaccurate Top-k𝑘kitalic_k predictions might help participants in identifying the correct label.

4.4. Willingness-to-Pay

We present the distributions of elicited willingness-to-pay (WTP) measures in Figure11 with summary statistics in Table10 of AppendixB.3. Elicited WTPs across conditions are quite similar, suggesting that the measure may be noisy. Participants are willing to pay slightly more for Top-10 ($1.84currency-dollar1.84\$1.84$ 1.84; CI: [1.58,2.11]1.582.11[1.58,2.11][ 1.58 , 2.11 ]) than RAPS ($1.78currency-dollar1.78\$1.78$ 1.78; CI: [1.51,2.07]1.512.07[1.51,2.07][ 1.51 , 2.07 ]) and Top-1 ($1.7currency-dollar1.7\$1.7$ 1.7; CI: [1.46,1.96]1.461.96[1.46,1.96][ 1.46 , 1.96 ]).

Nonetheless, as detailed in Section3.6, we quantify participants’ perceived utility for each prediction display using a measure we devised, called “willingness-to-overpay”. This proxy focuses on economic valuation, offering direct monetary comparisons between the monetary value of a prediction display that participants perceive (i.e., WTP) against how much additional expected bonus participants gained from accessing each prediction display.

ConditionAccuracy
Expected
Bonus
Expected
Bonus Diff.
baseline0.411.640.00
Top-10.632.520.86
Top-100.642.560.88
RAPS0.662.640.98

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (8)

\Description

This figure presents the bootstrapped 95% confidence intervals for the willingness-to-overpay measure by treatment conditions: Top-1, Top-10, and RAPS.

We note that this measure is imperfect as a comparator for any individual WTP because we cannot observe the counterfactual performance without a prediction display for each participant individually. In aggregate, however, Figure8 suggests that participants may over-value the prediction displays to a similar extent across conditions. If anything, despite benefiting more from using the prediction sets, we see weak evidence that RAPS participants perceive a lower monetary value from the prediction set relative to Top-k𝑘kitalic_k predictions. RAPS participants are willing to overpay by $0.8currency-dollar0.8\$0.8$ 0.8 on average (CI [0.54,1.1]0.541.1[0.54,1.1][ 0.54 , 1.1 ]), slightly less than Top-1 participants ($0.84currency-dollar0.84\$0.84$ 0.84, CI [0.598,1.10]0.5981.10[0.598,1.10][ 0.598 , 1.10 ]). Participants find Top-10 prediction most useful on average ($0.96currency-dollar0.96\$0.96$ 0.96; CI [0.7,1.23]0.71.23[0.7,1.23][ 0.7 , 1.23 ]).

4.5. Qualitative Analysis of Strategies

From our open coding, we find that participants employed different strategies between prediction displays (i.e., Figure10 in AppendixB.1). In general, most participants reported consulting predictions when decision-making (\approx81%percent8181\%81 % out of 450). Some participants reported first inspecting predictions and assessing their accuracy using their judgment (Top-1: \approx17%percent1717\%17 % out of 150; Top-10: 32%percent3232\%32 % out of 150; RAPS: \approx39%percent3939\%39 % out of 150), while others report relying first on their own intuitions to form a general idea, and then seeing if it aligned with any predictions (Top-1: \approx23%percent2323\%23 % out of 150; Top-10: 20%percent2020\%20 % out of 150; RAPS: \approx23%percent2323\%23 % out of 150). Very few participants chose to rely completely on their own intuitions because they find predictions inaccurate or not useful (Top-1: \approx3%percent33\%3 % out of 150; Top-10: \approx1%percent11\%1 % out of 150; RAPS: \approx1%percent11\%1 % out of 150).Others reported they would only consult predictions when the image was too hard to recognize and they had no clue (Top-1: 8%percent88\%8 % out of 150; Top-10: \approx25%percent2525\%25 % out of 150; RAPS: 10%percent1010\%10 % out of 150). Among participants who reported that they trusted AI prediction(s), some of them in Top-k𝑘kitalic_k conditions placed higher trust on predictions with high softmax pseudo-probability (e.g., 50%absentpercent50\geq 50\%≥ 50 %) (Top-1: \approx30%percent3030\%30 % out of 23; Top-10: \approx54%percent5454\%54 % out of 35), while those in RAPS condition focused more on the top suggestions in the prediction set (50%percent5050\%50 % out of 26).

Collectively participants used all features of our interface, reporting that the dropdown (\approx1%percent11\%1 % out of 599), bottom-up (\approx9%percent99\%9 % out of 599), and keyword search (\approx13%percent1313\%13 % out of 599) methods were useful in helping them identify a label, and the representative images were informative for aiding understanding of the meaning of labels (\approx14%percent1414\%14 % out of 599). Of participants who viewed prediction displays, those in the Top-1 condition appeared most likely to use the label space hierarchy network to find a better match (60% out of 150, compared to \boldsymbol{\approx}bold_≈23% out of 150 and 28% out of 150).

In addition to our open codes, we summarized the proportion of participants who fully relied on predictions (i.e., with all submitted responses selected from predictions). Aligning with the codes described above, we find Top-1 participants relied less on the predictions and used the search tool to find their preferred choice more often for both in-distribution (\approx39%percent3939\%39 % out of 150) and OOD (\approx3%percent33\%3 % out of 150) images. Top-10 participants reported relying heavily on predictions for in-distribution images (\approx95%percent9595\%95 % out of 150) and much less for OOD (25%percent2525\%25 % out of 150). RAPS participants similarly reported very frequently basing their decisions on a constrained set of predictions in-distribution (90%percent9090\%90 % out of 150) and slightly less but still quite often for OOD (48%percent4848\%48 % out of 150).

5. Discussion

Conformal prediction has gained recent visibility as a distribution-free approach for quantifying uncertainty in model predictions. Our results suggest that decision-makers may be able to use prediction sets to their advantage in an AI-advised labeling task, but that their advantages over status quo Top-k𝑘kitalic_k presentations vary with the properties of the setting. When the model has high accuracy and the test instances are in-distribution, we find that the size of the prediction set is a critical determinant of its utility in our setting. Smaller sets lead to performance on par with or slightly better than Top-k𝑘kitalic_k displays while larger ones lead to slightly worse performance. In contrast, when the model encounters unforeseen and challenging OOD instances that can cause Top-k𝑘kitalic_k predictions and their associated uncertainty scores to be unreliable (and which are also challenging for the humans), prediction sets offer some comparative advantage OOD, regardless of their set size, at least in a setting like ours where coverage guarantees are maintained.However, these advantages are relatively small, and may not always outweigh the benefits of instead using the calibration set instances in model training to further improve the classifier.We elucidate our main findings and discuss the broader implications of our study.

A smaller prediction set derived from a well-calibrated model is generally more useful.

Our results suggest that the utility of conformal prediction sets is linked to set size, presumably due to an underlying factor of cognitive load. A large set that embodies more uncertainty requires users to navigate and evaluate many incorrect predictions before finding the correct answer. This increase in cognitive load generally makes prediction sets less effective and may encourage more satisficing(Simon, 1956), resulting in a lower response accuracy. More rigorous uncertainty quantification does not always lead to better decision-making. Designers of prediction displays should carefully consider the trade-off between set size and adaptiveness when using conformal prediction sets to communicate prediction uncertainty for machine learning models in practical settings.

Conformal prediction sets can be more useful for some hard OOD instances.

One possible reason prediction sets can exceed in performance for hard OOD stimuli is their adaptiveness. While the covariate shifts used in our experimental setting can severely affect the accuracy of Top-k𝑘kitalic_k predictions, adaptive prediction sets under certain shifts can retain coverage by significantly increasing the set size (e.g., (Gendler etal., 2022)). While the high coverage of prediction sets for OOD instances does not necessarily equate to the high accuracy of AI-advised humans, as the results for hard instances inFigure6 demonstrate, even large set sizes for OOD instances provided some value to participants. Even if participants did not select the correct label from the set, when given larger prediction sets for hard OOD stimuli, they tended to select labels closer to the correct one, as measured by the shortest path length in the label space hierarchy. It is possible that the additional information about relative uncertainty provided by set size played a role in this advantage.

Withholding predictions of an uncalibrated model may improve decision quality

Consistent with prior work on AI-advised decision-making (e.g., (Green and Chen, 2019; Beede etal., 2020; Carton etal., 2020; Liu etal., 2021)), our results suggest that when a model is well-calibrated and more accurate than humans alone, users with access to its predictions can perform better than without the model, but not as well as the model alone for easier instances. When the model is poorly calibrated, display type affects whether people can perform better by accessing the model predictions. For instance, when the Top-k𝑘kitalic_k displays achieved lower coverage of the true label than the conformal guarantee, users sometimes performed worse than they would have done by ignoring the display. We observe qualitative feedback, such as “I used the Top-1111 when it was impossible to find what the image was based on my own perception” and “In cases where I really was unsure and did not know what the picture was, I relied on the AI and went with their label.” Decision-making quality, when presented with a constrained set of choices, is correlated with the calibration of prediction information. Only those using RAPS matched the expected performance of having no prediction display in hard OOD instances with larger set sizes. Our results underscore the potential value of treating the detection of instance type as a model’s learning problem (e.g., (Kaur etal., 2023)).

5.1. Limitations and Future Work

It is unclear whether a larger prediction set showing more possible labels can still be effective if the coverage is reduced. A natural follow-up study is to evaluate the efficacy of prediction sets in an OOD setting where coverage is disrupted at varying magnitudes, as well as one where the human retains their relative judgment accuracy and only the AI accuracy is affected. The adaptiveness of the prediction sets(Romano etal., 2020; Angelopoulos etal., 2020) enhances a model’s resilience to random and common perturbations that can occur in real-world scenarios, such as image corruptions. However, future work may explore the use of intentional adversarial methods that can attack softmax (e.g., FGSM(Goodfellow etal., 2015) or PGD(Madry etal., 2017)). Combined with the finding fromStraitouri and Rodriguez (2023) that constrained choices can lead to better performance based on smaller sets derived from a calibrated model, we might see an opposite effect to what we observed, where people are less likely to pick a choice from a prediction set that has a lower chance of containing the true label. While our experiment isolated the effect of the conformal prediction set and guarantee without performing other interpretability manipulations to the Top-k𝑘kitalic_k conditions, future work might also consider comparing conformal prediction sets to post-hoc calibrated softmax with Top-k𝑘kitalic_k displays.

Our experimental design, which does not observe individual participant performance without a prediction display, makes willingness-to-pay an imperfect comparator. Future work should delve into how users perceive the value of prediction displays and conduct within-subject comparisons of both performance and perceived value. Additionally, it is unclear whether the conformal coverage guarantee is prone to misconceptions common to conventional confidence intervals, such as misinterpreting the coverage rate and disregarding labels not included within the set.

6. Conclusion

Uncertainty quantification for deep neural networks (NNs) has been challenging due to their black-box nature. We evaluated whether communicating uncertainty via conformal prediction sets can improve human decision-making in AI-advised image labeling. We contribute the results of a large online experiment, in which we evaluate the utility of conformal prediction sets against status quo Top-k𝑘kitalic_k predictions across a diverse range of stimuli, varied by in-distribution or OOD, level of difficulty, and set size. Our work shows that when the model is well-calibrated, a smaller prediction set is most beneficial, as larger sets can overwhelm users and decrease performance due to increased cognitive load. However, when the model faces unexpected OOD instances that are also more difficult for humans, larger prediction sets can be more useful than their Top-k𝑘kitalic_k counterparts, at least when coverage remains high. Our work sheds light on how conformal prediction sets can be used to rigorously quantify uncertainty for machine learning models in practical applications and outlines potential avenues for future research.

References

  • (1)
  • Ambuehl and Li (2018)Sandro Ambuehl and Shengwu Li. 2018.Belief updating and the demand for information.Games and Economic Behavior 109 (2018), 21–39.
  • Amodei etal. (2016)Dario Amodei, Chris Olah, Jacob Steinhardt, PaulF. Christiano, John Schulman, and Dan Mané. 2016.Concrete problems in AI safety.arXiv preprint arXiv:1606.06565 (2016).
  • Angelopoulos etal. (2020)Anastasios Angelopoulos, Stephen Bates, Jitendra Malik, and MichaelI Jordan. 2020.Uncertainty sets for image classifiers using conformal prediction.arXiv preprint arXiv:2009.14193 (2020).
  • Angelopoulos and Bates (2021)AnastasiosN Angelopoulos and Stephen Bates. 2021.A gentle introduction to conformal prediction and distribution-free uncertainty quantification.arXiv preprint arXiv:2107.07511 (2021).
  • Babbar etal. (2022)Varun Babbar, Umang Bhatt, and Adrian Weller. 2022.On the utility of prediction sets in human-AI teams. In Proceedings of the Thirty-First International Joint Conferenceon Artificial Intelligence (IJCAI-22). 2457–2463.
  • Bansal etal. (2021)Gagan Bansal, Tongshuang Wu, Joyce Zhou, Raymond Fok, Besmira Nushi, Ece Kamar, MarcoTulio Ribeiro, and Daniel Weld. 2021.Does the whole exceed its parts? the effect of AI explanations on complementary team performance. In Proceedings of the 2021 ACM CHI Conference on Human Factors in Computing Systems. ACM, 1–16.
  • Beede etal. (2020)Emma Beede, Elizabeth Baylor, Fred Hersch, Anna Iurchenko, Lauren Wilcox, Paisan Ruamviboonsuk, and LauraM Vardoulakis. 2020.A human-centered evaluation of a deep learning system deployed in clinics for the detection of diabetic retinopathy. In Proceedings of the 2020 ACM CHI Conference on Human Factors in Computing Systems. ACM, 1–12.
  • Beller etal. (2013)Johannes Beller, Matthias Heesen, and Mark Vollrath. 2013.Improving the driver–automation interaction: An approach using automation uncertainty.Human Factors 55, 6 (2013), 1130–1141.
  • Budanitsky and Hirst (2006)Alexander Budanitsky and Graeme Hirst. 2006.Evaluating WordNet-based measures of lexical semantic relatedness.Computational Linguistics 32, 1 (2006), 13–47.
  • Carton etal. (2020)Samuel Carton, Qiaozhu Mei, and Paul Resnick. 2020.Feature-based explanations don’t help people detect misclassifications of online toxicity. In Proc. of the intl AAAI conference on web and social media, Vol.14. AAAI Press, 95–106.
  • Cumming (2014)Geoff Cumming. 2014.The new statistics: Why and how.Psychological Science 25, 1 (2014), 7–29.
  • Cumming and Finch (2005)Geoff Cumming and Sue Finch. 2005.Inference by eye: Confidence intervals and how to read pictures of data.American Psychologist 60, 2 (2005), 170–180.
  • Deng etal. (2009)Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009.ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 248–255.
  • Ekambaram etal. (2017)Rajmadhan Ekambaram, DmitryB Goldgof, and LawrenceO Hall. 2017.Finding label noise examples in large scale datasets. In Proc. of the 2017 IEEE Intl Conference on Systems, Man, and Cybernetics (SMC). IEEE, 2420–2424.
  • Fellbaum (1998)Christiane Fellbaum. 1998.WordNet: An electronic lexical database.MIT Press.
  • Fujiyoshi etal. (2019)Hironobu Fujiyoshi, Tsubasa Hirakawa, and Takayoshi Yamash*ta. 2019.Deep learning-based image recognition for autonomous driving.IATSS research 43, 4 (2019), 244–252.
  • Gelman etal. (2020)Andrew Gelman, Aki Vehtari, Daniel Simpson, CharlesC Margossian, Bob Carpenter, Yuling Yao, Lauren Kennedy, Jonah Gabry, Paul-Christian Bürkner, and Martin Modrák. 2020.Bayesian workflow.arXiv preprint arXiv:2011.01808 (2020).
  • Gendler etal. (2022)Asaf Gendler, Tsui-Wei Weng, Luca Daniel, and Yaniv Romano. 2022.Adversarially robust conformal prediction. In Proceedings of the ICLR.
  • Gibbs etal. (2023)Isaac Gibbs, JohnJ Cherian, and EmmanuelJ Candès. 2023.Conformal Prediction With Conditional Guarantees.arXiv preprint arXiv:2305.12616 (2023).
  • Goodfellow etal. (2015)IanJ Goodfellow, Jonathon Shlens, and Christian Szegedy. 2015.Explaining and harnessing adversarial examples.Proceedings of the ICLR.
  • Green and Chen (2019)Ben Green and Yiling Chen. 2019.The principles and limits of algorithm-in-the-loop decision making.In Proceedings of the ACM on Human-Computer Interaction 3, CSCW (2019), 1–24.
  • Guo etal. (2017)Chuan Guo, Geoff Pleiss, Yu Sun, and KilianQ Weinberger. 2017.On calibration of modern neural networks. In International conference on machine learning. PMLR, International Conference on Machine Learning (ICML), 1321–1330.
  • Guo etal. (2024)Ziyang Guo, Yifan Wu, Jason Hartline, and Jessica Hullman. 2024.A Statistical Framework for Measuring AI Reliance.arXiv preprint arXiv:2401.15356 (2024).
  • He etal. (2016)Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016.Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 770–778.
  • Hemmer etal. (2021)Patrick Hemmer, Max Schemmer, Michael Vössing, and Niklas Kühl. 2021.Human-AI Complementarity in Hybrid Intelligence Systems: A Structured Literature Review.PACIS (2021), 78.
  • Hendrycks and Dietterich (2019)Dan Hendrycks and Thomas Dietterich. 2019.Benchmarking Neural Network Robustness to Common Corruptions and Perturbations.In Proc. of the ICLR.
  • Hore and Barber (2023)Rohan Hore and RinaFoygel Barber. 2023.Conformal prediction with local weights: randomization enables local guarantees.arXiv preprint arXiv:2310.07850 (2023).
  • Hullman etal. (2024)Jessica Hullman, Alex Kale, and Jason Hartline. 2024.Decision Theoretic Foundations for Experiments Evaluating Human Decisions.arXiv preprint arXiv:2401.15106 (2024).
  • Hullman etal. (2018)Jessica Hullman, Xiaoli Qiao, Michael Correll, Alex Kale, and Matthew Kay. 2018.In pursuit of error: A survey of uncertainty visualization evaluation.IEEE Transactions on Visualization and Computer Graphics 25, 1 (2018), 903–913.
  • Hullman etal. (2015)Jessica Hullman, Paul Resnick, and Eytan Adar. 2015.Hypothetical outcome plots outperform error bars and violin plots for inferences about reliability of variable ordering.PloS One 10, 11 (2015), e0142444.
  • Joslyn and LeClerc (2012)SusanL Joslyn and JaredE LeClerc. 2012.Uncertainty forecasts improve weather-related decisions and attenuate the effects of forecast error.Journal of Experimental Psychology: Applied 18, 1 (2012), 126–140.
  • Kaur etal. (2023)Ramneet Kaur, Kaustubh Sridhar, Sangdon Park, Yahan Yang, Susmit Jha, Anirban Roy, Oleg Sokolsky, and Insup Lee. 2023.CODiT: Conformal out-of-distribution Detection in time-series data for cyber-physical systems. In Proceedings of the ACM/IEEE 14th International Conference on Cyber-Physical Systems (with CPS-IoT Week 2023) (ICCPS ’23). ACM, 120–131.
  • Kay etal. (2016)Matthew Kay, Tara Kola, JessicaR Hullman, and SeanA Munson. 2016.When (ish) is my bus? User-centered visualizations of uncertainty in everyday, mobile predictive systems. In Proceedings of the 2016 ACM CHI Conference on Human Factors in Computing Systems. ACM, 5092–5103.
  • Krizhevsky (2009)Alex Krizhevsky. 2009.Learning multiple layers of features from tiny images.Master’s Thesis, University of Toronto (2009).
  • Lai etal. (2022)Vivian Lai, Samuel Carton, Rajat Bhatnagar, Q.Vera Liao, Yunfeng Zhang, and Chenhao Tan. 2022.Human-AI Collaboration via Conditional Delegation: A Case Study of Content Moderation. In Proceedings of the 2022 ACM CHI Conference on Human Factors in Computing Systems. ACM, Article 54, 1–18pages.
  • Lai and Tan (2019)Vivian Lai and Chenhao Tan. 2019.On human predictions with explanations and predictions of machine learning models: A case study on deception detection. In Proc. of the conference on fairness, accountability, and transparency. ACM, 29–38.
  • Levy etal. (2021)Ariel Levy, Monica Agrawal, Arvind Satyanarayan, and David Sontag. 2021.Assessing the impact of automated suggestions on decision making: Domain experts mediate model errors but take less initiative. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–13.
  • Liu etal. (2021)Han Liu, Vivian Lai, and Chenhao Tan. 2021.Understanding the effect of out-of-distribution examples and interactive explanations on human-AI decision making.In Proc. ACM Hum.-Comput. Interact. 5, CSCW2 (2021), 1–45.
  • Madry etal. (2017)Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. 2017.Towards deep learning models resistant to adversarial attacks.arXiv preprint arXiv:1706.06083 (2017).
  • Manski (2019)CharlesF Manski. 2019.Communicating uncertainty in policy analysis.In Proceedings of the National Academy of Sciences 116, 16 (2019), 7634–7641.
  • Michaelis etal. (2019)Claudio Michaelis, Benjamin Mitzkus, Robert Geirhos, Evgenia Rusak, Oliver Bringmann, AlexanderS Ecker, Matthias Bethge, and Wieland Brendel. 2019.Benchmarking robustness in object detection: Autonomous driving when winter is coming.arXiv preprint arXiv:1907.07484 (2019).
  • Miller (1995)GeorgeA Miller. 1995.WordNet: A lexical database for English.Commun. ACM 38, 11 (1995), 39–41.
  • Nixon etal. (2019)Jeremy Nixon, MichaelW Dusenberry, Linchuan Zhang, Ghassen Jerfel, and Dustin Tran. 2019.Measuring calibration in deep learning. In Proceedings of the CVPR workshops, Vol.2. IEEE.
  • Northcutt etal. (2021)CurtisG Northcutt, Anish Athalye, and Jonas Mueller. 2021.Pervasive label errors in test sets destabilize machine learning benchmarks.arXiv preprint arXiv:2103.14749 (2021).
  • Papadopoulos etal. (2002)Harris Papadopoulos, Kostas Proedrou, Volodya Vovk, and Alex Gammerman. 2002.Inductive confidence machines for regression. In Proceedings of the Machine Learning: ECML 2002: 13th European Conference on Machine Learning. Springer Berlin, Heidelberg, 345–356.
  • Paszke etal. (2019)Adam Paszke etal. 2019.PyTorch: An imperative style, high-performance deep learning library.Advances in Neural Information Processing Systems 32 (2019).
  • Platt (1999)John Platt. 1999.Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods.Advances in Large Margin Classifiers 10, 3 (1999), 61–74.
  • Prabhudesai etal. (2023)Snehal Prabhudesai, Leyao Yang, Sumit Asthana, Xun Huan, QVera Liao, and Nikola Banovic. 2023.Understanding Uncertainty: How Lay Decision-makers Perceive and Interpret Uncertainty in Human-AI Decision Making. In Proceedings of the 28th International Conference on Intelligent User Interfaces. ACM, 379–396.
  • Quinonero-Candela etal. (2008)Joaquin Quinonero-Candela, Masashi Sugiyama, Anton Schwaighofer, and NeilD Lawrence. 2008.Dataset shift in machine learning.Mit Press.
  • Razzak etal. (2018)MuhammadImran Razzak, Saeeda Naz, and Ahmad Zaib. 2018.Deep learning for medical image processing: Overview, challenges and the future.Classification in BioApps: Automation of Decision Making (2018), 323–350.
  • Resnik (1995)Philip Resnik. 1995.Using Information Content to Evaluate Semantic Similarity in a Taxonomy. In Proceedings of the 14th International Joint Conference on Artificial Intelligence - Volume 1 (IJCAI’95). Morgan Kaufmann Publishers Inc., 448–453.
  • Romano etal. (2020)Yaniv Romano, Matteo Sesia, and Emmanuel Candes. 2020.Classification with valid and adaptive coverage.Advances in Neural Information Processing Systems 33 (2020), 3581–3591.
  • Sager etal. (2021)Christoph Sager, Christian Janiesch, and Patrick Zschech. 2021.A survey of image labelling for computer vision applications.Journal of Business Analytics 4, 2 (2021), 91–110.
  • Simon (1956)HerbertA Simon. 1956.Rational choice and the structure of the environment.Psychological Review 63, 2 (1956), 129–138.
  • Steyvers etal. (2022)Mark Steyvers, Heliodoro Tejeda, Gavin Kerrigan, and Padhraic Smyth. 2022.Bayesian modeling of human–AI complementarity.In Proceedings of the National Academy of Sciences 119, 11 (2022), e2111547119.
  • Straitouri and Rodriguez (2023)Eleni Straitouri and ManuelGomez Rodriguez. 2023.Designing decision support systems using counterfactual prediction sets.arXiv preprint arXiv:2306.03928 (2023).
  • Taylor and Kuyatt (1994)BarryN Taylor and ChrisE Kuyatt. 1994.Guidelines for evaluating and expressing the uncertainty of NIST measurement results. Vol.1297.National Institute of Standards and Technology.
  • Torralba and Efros (2011)Antonio Torralba and AlexeiA Efros. 2011.Unbiased look at dataset bias. In Proceedings of the CVPR. IEEE, 1521–1528.
  • Tversky and Kahneman (1971)Amos Tversky and Daniel Kahneman. 1971.Belief in the law of small numbers.Psychological bulletin 76, 2 (1971), 105.
  • vander Wal etal. (2021)Douwe vander Wal, Iny Jhun, Israa Laklouk, Jeff Nirschl, Lara Richer, Rebecca Rojansky, Talent Theparee, Joshua Wheeler, Jörg Sander, Felix Feng, etal. 2021.Biological data annotation via a human-augmenting AI-based labeling system.NPJ Digital Medicine 4, 1 (2021), 145.
  • Vovk etal. (2005)Vladimir Vovk, Alexander Gammerman, and Glenn Shafer. 2005.Algorithmic learning in a random world. Vol.29.Springer.
  • Zagoruyko and Komodakis (2016)Sergey Zagoruyko and Nikos Komodakis. 2016.Wide Residual Networks. In Proceedings of the British Machine Vision Conference 2016. British Machine Vision Association, BMVA Press.
  • Zhang etal. (2021)Dongping Zhang, Eytan Adar, and Jessica Hullman. 2021.Visualizing uncertainty in probabilistic graphs with Network Hypothetical Outcome Plots (NetHOPs).IEEE Transactions on Visualization and Computer Graphics 28, 1 (2021), 443–453.
  • Zhang etal. (2020)Yunfeng Zhang, QVera Liao, and RachelKE Bellamy. 2020.Effect of confidence and explanation on accuracy and trust calibration in AI-assisted decision making. In Proceedings of the 2020 ACM Conference on Fairness, Accountability, and Transparency. ACM, 295–305.
  • Zhou etal. (2017)Jianlong Zhou, SyedZ Arshad, Simon Luo, and Fang Chen. 2017.Effects of uncertainty and cognitive load on user trust in predictive decision making. In Proceedings of the 16th IFIP TC 13 International Conference on Human-Computer Interaction — INTERACT 2017. Springer-Verlag, 23–39.

Appendix A Further Methodological Details

We present the distribution of the cross-entropy loss for stimuli that are in-distribution in Figure9. Table6 shows WRN’s Top-k𝑘kitalic_k prediction accuracy and conformal coverage for in-distribution stimuli and 15 types of covariate shifts. Table8 outlines the sampling correction for images with a deterministic set, while Table9 explains the open codes used to categorize participants’ strategies. Furthermore, Table7 compares the performance of the Human-AI team, humans working alone, and AI standalone accuracy.

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (9)

\Description

This figure illustrates the distribution of cross-entropy loss for in-distribution predictions. To facilitate comparison of the median cross-entropy values, we use two visual cues: an orange dotted line represents the median when including images with a set size of one, while a red solid line denotes the median after excluding these images.

CorruptionCoverage
Top-1Top-10RAPS
none0.820.980.96
brightness0.770.960.95
contrast0.760.960.95
fog0.720.930.94
jpeg compression0.70.930.94
pixelate0.660.90.92
motion blur0.570.830.88
impulse noise0.570.840.89
shot noise0.560.830.89
gaussian noise0.550.830.89
elastic transform0.520.780.84
defocus blur0.520.810.91
snow0.510.790.86
frost0.430.70.76
zoom blur0.380.660.73
glass blur0.290.550.72

ConditionCovariate ShiftAccuracy
Human-AI TeamAI BaselineHuman Baseline
Top-1in0.740.730.53
out0.590.520.38
Top-10in0.760.730.53
out0.590.520.38
RAPSin0.730.730.53
out0.630.520.38

CorruptionDifficultySize1 CountSizeN CountSize RatioSizeN KeepSize1 Included
nonehard53927940.19749144
noneeasy1342829624.5317788060
defocus blurhard33177500.0487037
defocus blureasy457124311.88385724
snowhard66473190.0982775
snoweasy496425541.94451877
frosthard99390730.11996109
frosteasy289615901.82249454
zoom blurhard72290130.08106685
zoom blureasy292519141.53235359
glass blurhard51699730.05110857
glass blureasy202513601.49331493
CategoryOpen CodesDescription
1UNINFORMATIVEContent that doesn’t provide any significant or relevant information
2HUMAN-AI COMPARISONParticipant analyzes differences and similarities between own’s decisions and AI predictions
3HUMAN KNOWLEDGEParticipant makes decisions based solely on own intuition, wisdom, or experience
4HUMAN KNOWLEDGE 1ST & AI ADVICE 2NDParticipant makes decision based on own’s intuition and experience first, supplemented by AI advice or suggestions
5AI ADVICE 1ST & HUMAN CONFIRM 2NDParticipant seeks advice from AI first and then confirms with own’s knowledge
6AI ADVICE WHEN HARDParticipant seeks advice from AI only and specifically in challenging or complex scenarios (e.g., blurry images, etc.)
7TRUST AI PREDICTION(S)Participant has confidence in the prediction made by AI
8TRUST AI (RANKING OR CONFIDENCE)Participant relies on AI’s ranked suggestions or its confidence level in predictions. Participant mentions looking only at top predictions in RAPS and Top-10, while in Top-1 referring to specific confidence percentages
9TOP PREDICTION FALLACYParticipant expects to see the true answer in top predictions, but they believe the top prediction is wrong
10DISAPPOINTED AT AIParticipant specifically mentions dissatisfaction with AI’s performance and does not trust AI’s prediction
11SEARCHParticipant performs a general search without specifying
12SEARCH (BY DROPDOWN)Participant types keywords in the input field and refers to the dropdown for matches
13SEARCH (BY KEYWORD)Participant clicks “Search” to view a hierarchy tree that highlights matching categories
14SEARCH (BY AI LABEL)Participant clicks an AI-predicted label for a bottom-up search
15SEARCH (HIERARCHY NETWORK)Participant expands or subtracts nodes/categories and navigates through the provided hierarchy network
16ELIMINATIONParticipant tries to narrow down their options (to search in) by elimination
17REPRESENTATIVE IMAGESParticipant compares the task image against the given label-representative images
18COLOR & SURROUNDINGSParticipant relies on color, shapes, sizes, and surrounding objects to figure out the correct label

Appendix B Additional Results

B.1. Qualitative Analysis of Strategies

Figure10 presents a trellis plot summarizing the qualitative results per identified open code. Each row presents a code category with columns showing a barchart summarizing counts of each unique code from participants’ strategies by treatment conditions.

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (10)

\Description

This figure presents a trellis plot summarizing the qualitative results per identified open code. Each row presents a code category with columns showing a barchart summarizing counts of each unique code from participants’ strategies by treatment conditions.

B.2. Shortest Path Length

The shortest path (SP) length quantifies the “incorrectness” of participants’ labeling choices relative to the correct label in the label space hierarchy network. Similar to accuracy, we analyze the variations of participants’ labeling errors by task image types. We report the median of expected predictions for each type of image stimuli with uncertainty quantified as 95% HDI, holding the trial effect constant at the mean.

B.2.1. In-distribution

At a high level, the SP length is, on average, smaller for easy images across conditions. When task images are hard, the SP length is greater for images with larger set sizes across conditions.

When task images are easy, the SP length does not appear to be affected by set size—the distributions of expected SP length closely overlay across conditions (cf. Figure7, A and B). After marginalizing by set sizes, we find that when participants have access to predictions, even when the provided label is wrong, the incorrect label appears to be closer to the correct label in the distance (HPI: [1.05,1.6]1.051.6[1.05,1.6][ 1.05 , 1.6 ]) than those of the baseline condition (HPI: [2.01,2.93]2.012.93[2.01,2.93][ 2.01 , 2.93 ]).

When task images are hard but with a smaller average set size of 8888, Figure7 (C) shows the SP length is smallest for RAPS condition with a median of 1.481.481.481.48 (HPI: [1.34,1.64]1.341.64[1.34,1.64][ 1.34 , 1.64 ]). We do not detect differences between the two Top-k𝑘kitalic_k conditions, as HPIs closely overlap (Top-1: 1.81.81.81.8 HPI: [1.63,1.99]1.631.99[1.63,1.99][ 1.63 , 1.99 ] and Top-10: 1.781.781.781.78 HPI: [1.57,2]1.572[1.57,2][ 1.57 , 2 ]). However, when without predictions, the baseline participants tend to submit more incorrect answers, reflected by the largest SP length of 2.442.442.442.44 (HPI: [2.13,2.84]2.132.84[2.13,2.84][ 2.13 , 2.84 ]).

For hard task images with a larger average set size of 51515151, all intervals in Figure7 (D) get wider than those in (C), indicating greater variations. Although SP length increased across conditions, we do not detect noticeable differences among conditions with predictions (HPI: [2.98,3.42]2.983.42[2.98,3.42][ 2.98 , 3.42 ]). Similarly, the baseline participants produce more inaccurate responses with a median SP length of 3.833.833.833.83 (HPI: [3.31,4.39]3.314.39[3.31,4.39][ 3.31 , 4.39 ]).

In summary, we find that the SP length mirrors the accuracy in that prediction display that yields higher accuracy, which tends to produce a lower SP length. Between accuracy and SP length, a noticeable difference can be detected from task images that are hard with a smaller set size: although participants in the Top-1 and the baseline conditions have similar labeling accuracy, the SP length of Top-1 participants is noticeably lower than that of the baseline. This implies that predictions can help elicit responses closer to the correct label, even if they are wrong. Otherwise, our SP results are mostly consistent with our accuracy results.

B.2.2. Out-of-distribution

For OOD task images, the SP length between easy and hard images becomes more contrasting. For easy images, because predictions are still reliable and accurate, we find that the Top-1 condition tends to produce the lowest SP length of 1.351.351.351.35 (HPI: [1.24,1.49]1.241.49[1.24,1.49][ 1.24 , 1.49 ]) than Top-10 (1.671.671.671.67; HPI: [1.5,1.9]1.51.9[1.5,1.9][ 1.5 , 1.9 ]) and RAPS (1.891.891.891.89; HPI: [1.6,2.32]1.62.32[1.6,2.32][ 1.6 , 2.32 ]), after marginalizing over smaller and larger set sizes (cf. Figure7, E and F).

For hard OOD task images, we see an apparent mirroring effect relative to accuracy. When image stimuli have a smaller average set size of 30303030 shown in Figure7 (G), Top-1 participants have the lowest accuracy rate and also provide labels that have the highest SP length of 6.226.226.226.22 (HPI: [5.77,6.68]5.776.68[5.77,6.68][ 5.77 , 6.68 ]). RAPS participants have the highest accuracy for this type of image and produce the lowest SP length of 4.234.234.234.23 (HPI: [3.84,4.65]3.844.65[3.84,4.65][ 3.84 , 4.65 ]). The Top-10 condition overlaps with RAPS with an SP length of 4.784.784.784.78 (HPI: [4.37,5.19]4.375.19[4.37,5.19][ 4.37 , 5.19 ]), and both Top-10 and RAPS produce lower SP lengths than Top-1 and the baseline condition (5.675.675.675.67; HPI: [5.22,6.1]5.226.1[5.22,6.1][ 5.22 , 6.1 ]).

Similarly, for hard OOD task images with a larger average set size of 91919191 shown in Figure7 (H), baseline and RAPS conditions that can help participants achieve the highest labeling accuracy, also support them to give answers closer to the correct label with a median SP length that ranges between 3.033.033.033.03 and 3.683.683.683.68, followed by Top-1 (4.44.44.44.4; HPI: [3.99,4.79]3.994.79[3.99,4.79][ 3.99 , 4.79 ]) and Top-10 conditions (4.994.994.994.99; HPI: [4.61,5.41]4.615.41[4.61,5.41][ 4.61 , 5.41 ]).

However, despite that participants of the prediction set have similar accuracy for hard OOD images with smaller and larger sets, we find SP length is smaller for images with larger set sizes. This implies that a larger set size is helpful for hard OOD images, guiding participants to labels closer to the ground truth.

B.3. Distribution of Willingness-to-Pay

Table10 presents the summary statistics of the elicited willingness-to-pay, while Figure11 visualizes the distributions using a dot plot.

ConditionMin.Q1Q2Q3Max.MeanSD
RAPS011.0281.781.74
Top-1011.0281.701.56
Top-10011.5281.841.67

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (11)

\Description

This figure presents the distribution of willingness-to-pay elicited from participants by treatment conditions.

Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling (2024)
Top Articles
Latest Posts
Article information

Author: Dr. Pierre Goyette

Last Updated:

Views: 5823

Rating: 5 / 5 (50 voted)

Reviews: 89% of readers found this page helpful

Author information

Name: Dr. Pierre Goyette

Birthday: 1998-01-29

Address: Apt. 611 3357 Yong Plain, West Audra, IL 70053

Phone: +5819954278378

Job: Construction Director

Hobby: Embroidery, Creative writing, Shopping, Driving, Stand-up comedy, Coffee roasting, Scrapbooking

Introduction: My name is Dr. Pierre Goyette, I am a enchanting, powerful, jolly, rich, graceful, colorful, zany person who loves writing and wants to share my knowledge and understanding with you.