In last week's post, I noted the existence of a cognitive bias known as naïve realism, which is the human tendency to assume (usually incorrectly) that our perceptions of the world are accurate and objective. Naïve realism can contribute to our tendency to push ahead with projects even after we've noticed risks that present significant threats to success. I called these risks very bad news (VBN) risks.
Discussions about how to deal with VBN risks typically involve two sets of participants. Representing the "Business" are people from units such as Sales, Marketing, Human Resources, users, and executives. I call them, collectively, the Business. Representing the "technologists" are the people who will be doing the actual work of producing what the Business needs and wants. They include people from units such as product engineering, Information Technology (IT), and business technology. I call them, collectively, Engineering/IT.
And in that latest post, I described how the Business sometimes accepts VBN risks by rejecting the assertions of Engineering/IT as a result of naïve realism. The probability of occurrence of this phenomenon is elevated when the tactics available for addressing the VBN risk — avoidance, compensation, and mitigation — are so costly that they render the project unaffordable and a candidate for cancellation. So instead of abandoning the project because of its VBN risks, organizations tend to accept the risks more often than objectivity would allow.
Last week I closed with a question: What patterns of thought and decision making enable the group to accept a VBN risk, despite the stark disagreement between the Business and Engineering/IT?
In what follows I use the name Ben when referring to someone from the Business side of the decision, and the name Emma when referring to someone on the Engineering/IT side of the decision.
For concreteness, here's a description of a risk-acceptance situation:
Ben has been advocating for the Marigold project for over two years. Last year, he was able to secure some minimal funding for a study of the opportunity and some preliminary planning. That work led to approval for significant resources for this year. About a month into the work, Emma and her team found a VBN risk that they couldn't address with allocated resources. The risk they found was somewhat technical, but the Engineering/IT team felt that they had devised an explanation that the Business team would find persuasive. Their explanation showed why addressing the VBN risk would have undermined the business case for Marigold, at least from the financial perspective. This led Emma to have the Engineering/IT team develop a presentation for Ben and his team, including a case for Marigold's cancellation. Ben and Emma convened a meeting of everyone involved in Marigold to consider what to do. In the end, they all decided to push forward, accepting the VBN risk without a risk plan of any kind.
Within a few months, the VBN risk materialized and after several attempts to recover from it, Marigold was abandoned.
How did this happen? What enabled the team to set aside their disagreements about the VBN risk and push forward on a doomed project? What follows is an exploration of the phenomena that move such groups as they try to reach a go-no-go decision about a project for which they've identified a VBN risk.
In the remainder of this post, I describe one pathway through cognitive biases, other psychological phenomena, and organizational politics that leads to inappropriate risk acceptance. There are probably dozens of such paths, involving different psychological phenomena and organizational dysfunctions. This is only an example.
The Dunning-Kruger effect leads to overconfidence of the Business team
Justin Kruger and David Dunning, working at Cornell in 1999, discovered a phenomenon now called the Dunning-Kruger effect. [Kruger 1999] They performed experiments that yielded results consistent with the following four principles (paraphrasing):
- Incompetent individuals, compared to their more competent peers, dramatically overestimate their ability and performance.
- Incompetent individuals are less able than their more competent peers to recognize competence when they see it.
- Incompetent individuals are less able than their more competent peers to gain insight into the true level of their own performance.
- Incompetent individuals can gain insight about their shortcomings, but this comes (paradoxically) by gaining competence.
Because of the Dunning-Kruger effect, the members of the Business team tended to overestimate their own competence relative to handling the VBN risk. And they were also unable to appreciate the competence of Emma's team in assessing the severity of the risk. This gap in the levels of competence of the Business and Engineering/IT can create a gap in understanding that forms the basis of the risk acceptance that occurs in this case. And the level of competence of the Business team relative to managing the VBN risk can lead to errors as they assess their own competence. Those errors manifest themselves as overconfidence about how well they can deal with the VBN risk.
Motivated reasoning creates a gap in understanding
In the case Motivated reasoning can conspire with
the Dunning-Kruger Effect to cause
people to confidently adopt positions
that place the project at riskdescribed above, Ben and the other members of the Business team were very motivated to continue with the Marigold project. This motivation led to a biased assessment of the consequences of the VBN risk. 200819 [Kunda 1990] [Molden 2005] Because of motivated reasoning, the people representing the Business tend to underestimate the seriousness of the VBN risk. Emma and her team might also be engaged in motivated reasoning. But for the engineers, the bias leads them to tend to overestimate the seriousness of the VBN risk.
Motivated reasoning thus tends to create a sense of confidence among Ben's team that could become overconfidence. This situation is inherently asymmetric. The Business tends to adopt a risk-aggressive position; Engineering/IT tends to adopt a risk-averse position. Moreover, the two parties are subject to very different consequences if Marigold is cancelled. Engineering/IT almost certainly has a backlog of other work to do, which they are happy to tackle. But the Business will have to find another business opportunity. And since Marigold was probably the most attractive opportunity within reach, the Business will likely experience more disappointment following Marigold's cancellation than will Engineering/IT.
The power differential favors the Business
In many organizations, relative to questions of organizational direction, the power distance between the Business and Engineering/IT is significant. In organizations in which the Business is more powerful than Engineering/IT, the Business has greater influence over the organizational agenda. In such organizations, although it is the task of Engineering/IT to determine how to achieve the objectives of the Business, Engineering/IT has somewhat less influence relative to the question of whether to achieve those objectives.
The power differential thus accounts for some of the inability of Engineering/IT to stop Marigold on account of the VBN risk. But raw power, in the end, is costly to the organization. Its exercise is demoralizing for the people who must bow to it. For that reason, most groups search for reasons to support and comply with the preferences of the powerful. Such reasons provide a basis for the powerful to believe that they have considered all views and have led the group to a mutual accommodation; and they provide a basis for the less powerful to feel heard and to maintain professional dignity.
Overconfidence is contagious
When Enron entered into bankruptcy on December 3, 2001, it was the seventh-largest company in the United States. Subsequent research has determined that the organization had a "culture of arrogance," and that Enron employees regarded themselves as working in an elite organization, and smarter than anyone else. These beliefs account in part for Enron's pattern of negotiating deals that others would have regarded as risky to the point of being foolhardy.
Joey T. Cheng and colleagues have conducted experiments that suggest that in social groups, overconfidence can be transmitted from one person to the next, eventually taking root in the culture of the group. [Cheng 2020] [Cheng 2021] The phenomenon of overconfidence contagion could explain why the people representing the Business so uniformly regard a VBN risk as acceptable, saying, for example, "We'll cross that bridge if we come to it." While that uniformity of overconfidence can be remarkable, it can also arise from other factors, such as similarity of experience, education, or interests.
More intriguing is the possibility that overconfidence can be transmitted from the members of the Business team to members of the Engineering/IT team. During negotiations and working meetings, confronted with uniformly confident statements by Business team members, Engineering/IT team members "adjust" their views of the seriousness of the VBN risk. The risk of overconfidence contagion is especially elevated when the issue at hand is a VBN risk, because the significance of the issue is so strongly dependent on judgment and experience. Probabilities and consequences aren't physical things. Although we represent them as numbers, they aren't measurements. To assess them, we must rely on estimates and judgments. They are vulnerable to the effects of overconfidence contagion.
Let me review the path we've just taken.
- The Dunning-Kruger effect leads to overconfidence of Business teams with respect to their ability to tolerate VBN risks.
- Motivated reasoning creates a gap between the Business and Engineering/IT relative to their understanding of the nature and consequences of the VBN risk.
- The power differential between the Business and Engineering/It favors the Business, as they seek to have the organization adopt their approach to the VBN risk.
- The overconfidence of the Business team is contagious, which spreads through the Business team and unifies them. It also affects Engineering/IT and weakens their reticence to accept the VBN risk.
This is just one path that leads to risk acceptance, overcoming the cautious position of Engineering/IT and causing the organization to accept A VBN risk that should have been reason enough to halt the project. You might have encountered similar paths in your organization. Beware. A path once trodden is easier to follow a second time. First in this series Top Next Issue
Are you fed up with tense, explosive meetings? Are you or a colleague the target of a bully? Destructive conflict can ruin organizations. But if we believe that all conflict is destructive, and that we can somehow eliminate conflict, or that conflict is an enemy of productivity, then we're in conflict with Conflict itself. Read 101 Tips for Managing Conflict to learn how to make peace with conflict and make it an organizational asset. Order Now!
For more about the Dunning-Kruger Effect, see "The Paradox of Confidence," Point Lookout for January 7, 2009; "How to Reject Expert Opinion: II," Point Lookout for January 4, 2012; "Devious Political Tactics: More from the Field Manual," Point Lookout for August 29, 2012; "Overconfidence at Work," Point Lookout for April 15, 2015; "Wishful Thinking and Perception: II," Point Lookout for November 4, 2015; "Wishful Significance: II," Point Lookout for December 23, 2015; "Cognitive Biases and Influence: I," Point Lookout for July 6, 2016; and "The Paradox of Carefully Chosen Words," Point Lookout for November 16, 2016.
Your comments are welcomeWould you like to see your comments posted here? rbrenIyeJIiAfnGdKlUXrner@ChacsxirZwZlENmHUNHioCanyon.comSend me your comments by email, or by Web form.
About Point Lookout
Thank you for reading this article. I hope you enjoyed it and found it useful, and that you'll consider recommending it to a friend.
This article in its entirety was written by a human being. No machine intelligence was involved in any way.
Support Point Lookout by joining the Friends of Point Lookout, as an individual or as an organization.
Do you face a complex interpersonal situation? Send it in, anonymously if you like, and I'll give you my two cents.
More articles on Cognitive Biases at Work:
- Wishful Significance: I
- When things don't work out, and we investigate why, we sometimes attribute our misfortune to "wishful
thinking." In this part of our exploration of wishful thinking we examine how we arrive at mistaken
assessments of the significance of what we see, hear, or learn.
- The Rhyme-as-Reason Effect
- When we speak or write, the phrases we use have both form and meaning. Although we usually think of
form and meaning as distinct, humans tend to assess as more meaningful and valid those phrases that
are more beautifully formed. The rhyme-as-reason effect causes us to confuse the validity of a phrase
with its aesthetics.
- On Standing Aside
- Occasionally we're asked to participate in deliberations about issues relating to our work responsibilities.
Usually we respond in good faith. And sometimes we — or those around us — can't be certain
that we're responding in good faith. In those situations, we must stand aside.
- Motivated Reasoning and the Pseudocertainty Effect
- When we have a preconceived notion of what conclusion a decision process should produce, we sometimes
engage in "motivated reasoning" to ensure that we get the result we want. That's risky enough
as it is. But when we do this in relation to a chain of decisions in the context of uncertainty, trouble
- Clouted Thinking
- When we say that people have "clout" we mean that they have more organizational power or social
influence than most others do. But when people with clout try to use it in realms beyond those in which
they've earned it, trouble looms.
Forthcoming issues of Point Lookout
- Coming February 28: Checklists: Conventional or Auditable
- Checklists help us remember the steps of complex procedures, and the order in which we must execute them. The simplest form is the conventional checklist. But when we need a record of what we've done, we need an auditable checklist. Available here and by RSS on February 28.
- And on March 6: Six More Insights About Workplace Bullying
- Some of the lore about dealing with bullies at work isn't just wrong — it's harmful. It's harmful in the sense that applying it intensifies the bullying. Here are six insights that might help when devising strategies for dealing with bullies at work. Example: Letting yourself be bullied is not a thing. Available here and by RSS on March 6.
I offer email and telephone coaching at both corporate and individual rates. Contact Rick for details at rbrenIyeJIiAfnGdKlUXrner@ChacsxirZwZlENmHUNHioCanyon.com or (650) 787-6475, or toll-free in the continental US at (866) 378-5470.
Get the ebook!
Past issues of Point Lookout are available in six ebooks:
- Get 2001-2 in Geese Don't Land on Twigs (PDF, )
- Get 2003-4 in Why Dogs Wag (PDF, )
- Get 2005-6 in Loopy Things We Do (PDF, )
- Get 2007-8 in Things We Believe That Maybe Aren't So True (PDF, )
- Get 2009-10 in The Questions Not Asked (PDF, )
- Get all of the first twelve years (2001-2012) in The Collected Issues of Point Lookout (PDF, )
Are you a writer, editor or publisher on deadline? Are you looking for an article that will get people talking and get compliments flying your way? You can have 500-1000 words in your inbox in one hour. License any article from this Web site. More info