[2/3] Believing or doubting: the issue of confidence bias in decision-making

Rédigé par Charlotte Barot

 - 

17 October 2024


As decision-support systems become increasingly widespread, their use in critical decision-making contexts raises serious ethical and legal challenges. To address the main risks identified in the decision-making process, the law requires human intervention or integrated human oversight within the procedure, resulting in “hybrid” mechanisms that combine computational power with human judgment. In this series of articles, the LINC examines, based on scientific literature, two key obstacles to the effectiveness of such mechanisms : on the one hand, user trust biases toward the system, and on the other, the opacity of the system’s suggestions.

 

This article is the second in a series of three :

In December 2023, the Court of Justice of the European Union ruled that the credit scoring tool of the German company SCHUFA, which provided banks with an estimate of a client’s creditworthiness in the form of a trust score, constituted a fully automated decision. In practice, however, the decision to grant credit was formally made by a bank employee responsible for verifying and either applying or rejecting the tool’s proposal, thereby implying human intervention. Yet, since the score was never actually contested by bank employees, who systematically relied on it, effectively turning the suggestion into the credit decision, the Court concluded that such intervention was not meaningful.

To be more than a mere formality, human oversight, as defined in the European AI Act, must make it possible to detect errors, diverge from, or interrupt the system ; in other words, to provide a genuine alternative to the AI system’s output. These abilities, however, rely on psychological dispositions highlighted by the regulation, which specifies that the person responsible for oversight must be aware of potential cognitive biases, such as excessive trust. This bias poses a risk of undermining the quality of the oversight exercised, and consequently, the reliability of the entire mechanism.

Empirical literature does not point to a uniform reaction of individuals toward algorithms but instead identifies two tendencies : either an acceptance heuristic (the appreciation thesis), noted in the European AI Act, or a rejection heuristic (the aversion thesis), despite the errors that both strategies can produce.

 

Too much trust brings risk: excessive appreciation

 

Some studies highlight a tendency among participants to conform to the system’s outputs (e.g. Jacobs et al. 2021Green 2019Yin 2019Bussone et al. 2015Kiani et al. 2020Alberdi et al. 2004Logg et al. 2019Robinette et al. 2017). In such cases, human oversight ceases to be effective, as the decision-maker systematically relies on the system’s output, mirroring the logic of the Court of Justice of the European Union’s ruling on SCHUFA. Here, it is not the effectiveness of the decision itself that is in question, but rather the effectiveness of human oversight over algorithmic decision-making, and its ability to reject potential system errors or anomalies.

In a study conducted in psychiatry (Jacobs et al. 2021), researchers asked a cohort of clinicians to make a series of decisions about a fictitious patient. For each patient, the clinician had to decide on a treatment, either with a recommendation from a machine learning system accompanied by a brief justification of the suggested choice, or entirely without any recommendation (a completely independent decision).

The study shows that, on the one hand, the performance of the groups with and without access to the machine learning system’s recommendations is virtually the same, and that both groups make poorer decisions than the system alone. On the other hand, when the system produces an incorrect decision (where the error is defined as a position diverging from that of a panel of psychiatry experts), performance drops compared to the control group (without access to the machine learning system) that is, humans tend to conform to the algorithm’s decision in such cases.

Finally, the authors observed an effect of familiarity with the tool : clinicians who were more familiar with the system were, on average, less likely to follow a machine learning system’s recommendation, regardless of its accuracy, compared to clinicians who were less familiar with such systems.

The observation that the level of professional expertise plays a role in decision-making is corroborated by a study by Gaube et al. 2021 in which two groups of doctors with different levels of medical expertise were asked to produce a diagnosis and rate a recommendation displayed as coming from either an algorithm or a human, when in fact all the recommendations came from a human.

Doctors in the most expert group tend to rate recommendations less highly when they are indicated as coming from an AI system. However, the quality of their diagnosis is influenced by the quality of the advice received, regardless of its source (AI or human).

This result suggests that the influential effects of the system's advice could be a simple artifact of not having been able to form a decision before the confrontation, and not necessarily due to an attitude of deference toward the algorithm. Indeed, the study shows that participants generally tended to follow the advice given, whether it came from a human or a machine.

Absence of deference to the system by experts has also been observed in other studies (Logg et al. 2020Povyakalo et al. 2013). In the latter study, assessment was also modulated by disagreement. Advice that contradicted the participant's prior opinion had less impact but did not completely cancel out the effect of trust in the system. The assessment of algorithms decreased (but did not disappear) when their advice was contrary to their own judgment. The authors conclude that it is on these points of disagreement that people are most likely to improve their accuracy. These critical cases of confrontation are therefore interesting for decision-making.

 

Aversion and excessive mistrust

 

Some studies suggest excessive mistrust of the system, preventing participants from making informed decisions (e.g. Dietvorst et al. 2015Longoni et al. 2019Dzindolet et al. 2002Lim et O’Connor 1995Yeomans et al. 2019Promberger et Baron 2006).

In several forecasting tasks where they had to choose between an algorithmic prediction and a human prediction regarding the likelihood of a song's success (Dietvorst et al. 2015), when they saw the system working, and sometimes getting it wrong, participants tended to reject its predictions in favor of human advice, despite the higher error rate of human predictions (up to twice that of the algorithm).

In short, humans are less forgiving of algorithms' mistakes and tend to generalize their overall performance based on these harmful examples. This tendency persists even after observing that the system's performance exceeds that of humans on average. A natural explanation is that mistrust of algorithms exists even before observing them, and that observing mistakes reinforces this preconception.

There would therefore appear to be an anchoring bias at work: once users have formed an opinion about an AI system, it is very difficult for them to change their minds, even in the face of contradictory evidence. This is supported by the fact that aversion to algorithms is not observed in purely deterministic tasks such as logical calculations or memory tasks, where humans are notoriously more fallible than algorithms.

 

Amplification of biases and harmful interactions

 

Humans are not free from bias and do not have infallible judgment, which leads to errors that, all other things being equal, are added to those of the systems or amplify them.

 

The COMPAS case

 

The COMPAS  (Correctional Offender Management Profiling for Alternative Sanctions) system is a prediction tool used in the criminal justice system in the United States. Developed by Equivant (formerly Northpointe), this system assesses the risk that an offender will commit further offenses or fail to appear at a future hearing. COMPAS uses a set of questions about offenders' criminal history, behavior, and personal characteristics (social relationships, demographic data: age, gender, ethnicity, etc.) to calculate risk scores, including the risk of violent recidivism. Judges often use COMPAS scores to decide whether a defendant can be released pending trial. A 2016 investigation by ProPublica revealed that COMPAS had discriminatory biases, overestimating the risk of recidivism in certain individuals and leading to harsher sentences, without this being justified by criteria other than ethnicity.

To determine whether these biases stem solely from the algorithm Green et al. 2019 replicated the COMPAS system under laboratory conditions. The experimental task consisted of assessing the risk of recidivism of an individual, following the same logic as the original algorithm. Each individual’s description was accompanied by a recidivism risk score expressed as a percentage, and the participant had to indicate their own risk assessment. The researchers found that when an algorithmic suggestion was provided, participants made poorer decisions than the algorithm itself and were unable to evaluate either their own performance or that of the algorithm. They also emphasized that decision-makers were clearly biased against certain profiles, which led them to amplify the algorithm’s bias. Human biases, combined with high trust in the algorithm, can therefore result in not only accepting a biased outcome but also reinforcing its skew.

On this topic, see also the interviews of Angèle Christin (« Les méthodes ethnographiques nuancent l’idée d’une justice prédictive et entièrement automatisée ») and of Philippe Besse (« Les décisions algorithmiques ne sont pas plus objectives que les décisions humaines ») on the website of the LINC.

 

Appropriate judgment: influences and conditions for exercise

 

The observations outlined above highlight the need to pay particular attention to the conditions that enable the user to be in the best possible psychological state for exercising sound judgment. This concerns not only the user’s own skills and the task to be performed, but also a range of external factors related to the work environment and the context in which the decision is made. Among these, one can note :

 

The decision-making environment:

  • The cost of an error or being involved in a high-risk situation (e.g. Robinette and al. 2017). The risks assumed by the decision-maker, given the consequences for the affected individual, influence their judgment, as following or diverging from a system’s recommendation can lead to different outcomes in case of an error. In particular, the assignment of responsibility to the decision-maker constitutes a cost borne by that individual. If responsibility for a decision is attributed to the algorithm, it may seem costly for the human involved to diverge from the algorithm’s decision, which encourages them to accept the algorithm’s recommendations.
  • The time allocated for decision-making (e.g. Robinette et al. 2017). Very short deliberation periods lead to decisions that closely follow the algorithm’s suggestions, due to a reflexive effort to conserve cognitive resources.

 

Domain expertise:

  • The level of expertise in the given task (e.g. Jacobs et al. 2021Logg et al. 2019Povyakalo et al. 2013). Experts tend to rely less on the algorithm than novices, which is advantageous for the ability to diverge from the algorithm’s proposed decision, but it can also lead the expert toward potentially excessive confidence. Conversely, less experienced individuals are more likely to accept the system’s recommendation more frequently.
  • Doubt regarding the decision to be made. Cases in which the decision-maker is uncertain or experiences a high level of doubt are likely to be those in which the system’s suggestion has the greatest influence. This may occur in difficult situations due to an overwhelming number of options or because the case at hand is unique.

 

Relationship with automated systems:

  • Familiarity with algorithms and automated systems (e.g. Jacobs et al. 2021). Individuals who are less familiar with these systems tend, depending on the situation, either to exhibit greater trust or, conversely, excessive distrust, but their disposition is rarely neutral.
  • Prior trust in the system, i.e. preconceived notions about the algorithm’s reliability, which are very difficult to overcome, especially if they are based on observations that the system has made errors. (e.g. Robinette et al. 2017, Dietvorst et al. 2015, Prahl et al. 2017).
  • Congruence, or the intuitive agreement with the algorithm’s output (e.g. Logg et al. 2019). This seemingly trivial factor can strongly impact the human ability to question the output in cases of excessive distrust toward the algorithm.

 

Finally, the overall configuration in which the interaction between the decision-maker and the decision-support system takes place:

  • Anticipated suggestion. The system provides a recommendation even before the human makes a decision. The human then decides whether to accept the recommendation or to diverge and propose an alternative. As noted earlier, this configuration can encourage anchoring effects, making it more difficult for the human to diverge.
  • Doubt resolution. The system does not initiate any action but flags a situation in which it detects a potential risk. In this case, human intervention occurs only when an employee decides whether to validate the alert. Often used in remote monitoring, this configuration reduces the need for continuous human involvement, as personnel intervene only in certain situations, those indicating danger or requiring action. It requires that all potentially hazardous situations be accurately identified by the system.
  • Alternative suggestion. The system provides additional information to support a human decision that has already been considered. This approach helps prevent overreliance on the algorithm, but it can also reduce the effectiveness of its use, as the human operator has already made a decision and may find it “costly” to change it. However, in cases where the human is uncertain about their decision, this configuration appears to be particularly relevant.
  • Use of a choice algorithm. A choice algorithm determines whether the decision is returned to the human or handled solely by the system (a single, uncontested decision). This approach, proposed by Mozannar et al. 2023
  • Mozannar et al. 2023 helps avoid human trust biases, but it relies on another automated system, which does not meet the requirements of the GDPR and the AI Regulation, unless multiple layers of human intervention are implemented.
  • Index in a bundle. Independently, a human and an AI make a decision on a problem ; a third party, such as a panel or another expert, resolves any disagreements, ideally in a blind evaluation.

 

The «confrontation» options, namely options 3 and 5, where the user has the opportunity to form an independent judgment, appear best suited to allow the person not to simply “submit” to the system’s decisions, and thus to exercise their judgment. However, for these configurations to be effective, the human must be willing to genuinely compare their decision with the system’s suggestion, without necessarily trying to confirm their own initial hypothesis. While these two configurations are therefore relevant in the context of human oversight, the cases that are likely to be truly beneficial are those where the agent’s opinion is not already strongly fixed ; otherwise, there is a risk of simply rejecting the system’s decision.

 

Conclusion

 

Even with perfect neutrality, infallible judgment, and ideal working conditions, a fundamental obstacle remains: the very ability to decipher and make sense of the system’s suggestion. It is this intrinsic difficulty that we explore in the article " Predicting without explaining, or when algorithmic opacity muddies the waters ".