
Left to right, Deputy Secretary of Defense Paul Wolfowitz, Defense Secretary Donald H. Rumsfeld, and President George W. Bush, conduct a press conference on September 17, 2001, at the Pentagon. They had met to continue planning an appropriate military response to the al Qaeda threat. They held a press conference after the meeting. Certainly, over the months and years following, this team and many other members of the administration underestimated the scale of the problems they faced. Mr. Wolfowitz, in particular, expressed views far more optimistic than — and contrary to — the evidence-based views of many experts. One of his most often quoted comments about the cost of the Iraq War is that in the case of Iraq, the United States is dealing with "a country that can really finance its own reconstruction and relatively soon." Defense Department photo by R. D. Ward.
Overconfidence is the state of having too much confidence — confidence beyond levels justified by evidence. One trouble with that definition is that it provides little useful insight: how much is too much? A second trouble is that it exemplifies itself, in that it presumes that levels of confidence can be assigned with, um, confidence. Often we can do no such thing. Situations affected by overconfidence include hiring, making strategic choices, chartering projects, cancelling projects — indeed, most workplace decision making. Despite the vagueness of the concept of overconfidence we can make useful conclusions, if we examine the concept more closely.
That's what Don Moore and Paul Healy did in a 2008 paper — cited in over 800 other works (according to Google Scholar), which is a goodly number for such a short time. The authors note that conflicting results in overconfidence research can be resolved when one realizes that the term overconfidence had been used to denote three different classes of judgment errors. They are:
- Overestimation: assessing as too high one's actual ability, performance, level of control, or chance of success.
- Overplacement: the belief that one is better than others, such as when a majority of people rate themselves "better than average."
- Overprecision: excessive certainty regarding the accuracy of one's beliefs.
These tendencies are not character flaws. Rather, they arise from the state of being human — not in the sense of "to err is human," but, as a direct consequence of human psychology.
What is surprising is how little we do in organizations to protect ourselves and the organization from the effects of overconfidence. Indeed, some of our behaviors and policies actually induce overconfidence. Here are three examples.
- Unrealistic assessments of the capabilities of others
- A phenomenon known as the Dunning-Kruger Effect causes us to confuse competence and confidence. [Kruger 1999] That is, we assess people as more capable when they project confidence, and inversely, less capable when they project uncertainty. This can lead to decision-making errors in hiring, and in evaluating the advice we receive from subordinates, consultants, experts, and the media.
- Unrealistic standards of precision
- When we Situations affected by overconfidence
include hiring, making strategic choices,
chartering or cancelling projects —
indeed, most workplace decision makingevaluate projected performance for projects or business units, we require alignment between projections and actuals. The standards we apply when we assess performance typically exceed by far any reasonable expectations of the precision of those projections. This behavior encourages those making projections to commit the overprecision error. - Unrealistic risk appetite
- Assessments of success in the context of risk, and our ability to mitigate risk, are subject to overestimation errors. By overestimating our chances of success, and our ability to deal with adversity, we repeatedly subject ourselves to higher levels of risk than we realize.
Cognitive biases that contribute to overconfidence in its various forms include, among others, the planning fallacy, optimism bias, illusory superiority, and, of course, the overconfidence effect. Most important, the bias blind spot causes us to be overconfident about the question of whether we ourselves are ever overconfident. We surely are. At least, I think so. Top
Next Issue
Is every other day a tense, anxious, angry misery as you watch people around you, who couldn't even think their way through a game of Jacks, win at workplace politics and steal the credit and glory for just about everyone's best work including yours? Read 303 Secrets of Workplace Politics, filled with tips and techniques for succeeding in workplace politics. More info
For an extensive investigation of the role of overconfidence in governmental policies that lead to war, see Dominic D. P. Johnson, Overconfidence and War: The Havoc and Glory of Positive Illusions, Cambridge, Massachusetts: Harvard University Press, 2004. Order from Amazon.com.
For more about the Dunning-Kruger Effect, see "The Paradox of Confidence," Point Lookout for January 7, 2009; "How to Reject Expert Opinion: II," Point Lookout for January 4, 2012; "Devious Political Tactics: More from the Field Manual," Point Lookout for August 29, 2012; "Wishful Thinking and Perception: II," Point Lookout for November 4, 2015; "Wishful Significance: II," Point Lookout for December 23, 2015; "Cognitive Biases and Influence: I," Point Lookout for July 6, 2016; "The Paradox of Carefully Chosen Words," Point Lookout for November 16, 2016; and "Risk Acceptance: One Path," Point Lookout for March 3, 2021.
Footnotes
Your comments are welcome
Would you like to see your comments posted here? rbrenZLkFdSHmlHvCaSsuner@ChacbnsTPttsdDaRAswloCanyon.comSend me your comments by email, or by Web form.About Point Lookout
Thank you for reading this article. I hope you enjoyed it and found it useful, and that you'll consider recommending it to a friend.
Point Lookout is a free weekly email newsletter. Browse the archive of past issues. Subscribe for free.
Support Point Lookout by joining the Friends of Point Lookout, as an individual or as an organization.
Do you face a complex interpersonal situation? Send it in, anonymously if you like, and I'll give you my two cents.
Related articles
More articles on Personal, Team, and Organizational Effectiveness:
When Your Boss Is a Micromanager
- If your boss is a micromanager, your life can be a seemingly endless misery of humiliation and frustration.
Changing your boss is one possible solution, but it's unlikely to succeed. What you can do
is change the way you experience the micromanagement.
Commitment Makes It Easier
- When you face obstacles, sometimes the path around or through them is difficult. Committing yourself
to the path lets you focus all your energy on the path you've chosen.
Email Antics: I
- Nearly everyone I know complains that email is a time waster. Yet much of the problem results from our
own actions. If you're looking around for some New Year's resolutions to make, here are some ideas,
in this Part I of a little catalog of things we do that help waste our time.
Risk Management Risk: II
- Risk Management Risk is the risk that a particular risk management plan is deficient. Here are some
guidelines for reducing risk management risk arising from risk interactions and change.
The Tyranny of Singular Nouns
- When groups try to reach decisions, and the issue in question has a name that suggests a unitary concept,
such as "policy," they sometimes collectively assume that they're required to find a one-size-fits-all
solution. This assumption leads to poor decisions when one-size-fits-all isn't actually required.
See also Personal, Team, and Organizational Effectiveness, Critical Thinking at Work and Cognitive Biases at Work for more related articles.
Forthcoming issues of Point Lookout
Coming March 29: Time Slot Recycling: The Risks
- When we can't begin a meeting because some people haven't arrived, we sometimes cancel the meeting and hold a different one, with the people who are in attendance. It might seem like a good way to avoid wasting time, but there are risks. Available here and by RSS on March 29.
And on April 5: The Fallacy of Division
- Errors of reasoning are pervasive in everyday thought in most organizations. One of the more common errors is called the Fallacy of Division, in which we assume that attributes of a class apply to all members of that class. It leads to ridiculous results. Available here and by RSS on April 5.
Coaching services
I offer email and telephone coaching at both corporate and individual rates. Contact Rick for details at rbrenZLkFdSHmlHvCaSsuner@ChacbnsTPttsdDaRAswloCanyon.com or (650) 787-6475, or toll-free in the continental US at (866) 378-5470.
Get the ebook!
Past issues of Point Lookout are available in six ebooks:
- Get 2001-2 in Geese Don't Land on Twigs (PDF, )
- Get 2003-4 in Why Dogs Wag (PDF, )
- Get 2005-6 in Loopy Things We Do (PDF, )
- Get 2007-8 in Things We Believe That Maybe Aren't So True (PDF, )
- Get 2009-10 in The Questions Not Asked (PDF, )
- Get all of the first twelve years (2001-2012) in The Collected Issues of Point Lookout (PDF, )
Are you a writer, editor or publisher on deadline? Are you looking for an article that will get people talking and get compliments flying your way? You can have 500-1000 words in your inbox in one hour. License any article from this Web site. More info
Follow Rick





Recommend this issue to a friend
Send an email message to a friend
rbrenZLkFdSHmlHvCaSsuner@ChacbnsTPttsdDaRAswloCanyon.comSend a message to Rick
A Tip A Day feed
Point Lookout weekly feed
