From time to time, we become so involved in debating with each other that we can forget that for most issues we debate at work, Reality has the final say. It happens like this. We encounter a situation that has no obvious or immediate resolution. We debate the issue for a time, and finally settle on a path that nobody is really comfortable with. Then we push ahead, hoping that somehow it will all work out.
But Reality — not any debater's skill or stature — ultimately decides every question. Here's an example:In scenarios like this, people don't seem concerned that Reality hasn't had a chance to contribute to the debate. Nobody has suggested a "pilot," and the Pragmatists failed to sway the Visionaries by citing the difficulties previous efforts have had. And so, Reality is set aside until Franklin's Compromise Approach is either successful — or not. It's white-knuckle time.We go back and forth on this question, until someone who's highly respected — someone with "clout" whom I'll call Franklin — proposes a Compromise Approach. The Visionaries and Pragmatists agree to hire a few more people and acquire more powerful software. But although these measures were among those the Pragmatists wanted, much of what they wanted is excluded.
Clouted thinking
What The person who has the clout isn't necessarilyengaged in nefarious activity. It's often the
person with clout who is thinking most cloutedly.has happened in scenarios like the one above might be regarded as analogous to the consequences of a cognitive bias known as the halo effect. [Thorndike 1920] The halo effect is our tendency to allow positive (negative) impressions of one attribute of a person, company, country, brand, product, or any entity, really, to positively (negatively) influence our assessment of other attributes of that same person, company, country, brand, product, or entity. In the scenario above, both Visionaries and Pragmatists accept Franklin's Compromise Approach. It might be a good solution. But it's also possible that the halo effect has taken over. It's possible that their willingness to adopt Franklin's compromise was influenced, in part, by Franklin's stature — by his clout. Their thinking might have been something like, "Franklin knows his stuff, so maybe his idea will actually work." And that's the problem. Franklin knows his stuff. He has a record of accomplishment in some specific domains in which he earned his clout. But is that record applicable to the matter at hand? Perhaps. In many cases it is applicable. But too often, the halo effect causes us to accept the suggestions of the Franklins of the world even when the domains in which they earned their clout aren't relevant to the matter at hand. I call this pattern clouted thinking. The person who has the clout isn't necessarily engaged in nefarious activity. It's often the person with clout who is thinking most cloutedly.
Indicators of clouted thinking
Clouted thinking is one of the harmful effects of clout. Detecting clouted thinking can be difficult. But one aid in avoiding trouble is sensitivity to the kinds of comments people make when they're engaging in clouted thinking. Here are five indicators that the risk of clouted thinking is elevated.- Discarding some evidence while crediting other evidence
- Knowing what to attend to and what to set aside is essential for resolving complex problems. Be alert to the arbitrary application of the technique. Inconsistency in what we accept or reject can be driven by a desire to reach a specific outcome.
- I prefer my opinion to yours
- Preference for one opinion over others is not evidence of the validity of that opinion. Preference for an opinion is not a reason to be guided by that opinion. Demand evidence.
- If you can't explain why it's happening, it isn't happening
- It's irrational to reject an observation of system behavior as invalid on the basis that no known model of the system can account for that behavior. Instead, when our models cannot account for system behavior, we must accept that our models of the system are incomplete or incorrect. Models can be wrong. Reality is always right.
- It can't (must) be true because people with clout say so
- Accepting as facts the assertions of people with clout is one way of detaching the debate from Reality. The pronouncements of people with clout then become facts for purposes of the debate. And unlike actual facts, pronouncements can be mistaken or fabricated. Demand Reality-based evidence.
- You're wrong, because you're contradicting someone with clout
- People with clout are typically correct more often than others, but they do make mistakes, too. And people who lack clout can occasionally report correct observations, or occasionally have good ideas. It's irrational to reject (or accept) a statement on the sole basis of its author's stature. Demand evidence.
Last words
The suggestions above mention the need for evidence repeatedly, but they don't define the term. Evidence is fact. Statements by reliable individuals are sometimes the closest we come to gathering facts. Statements, when taken as fact, can sometime lead us to clouted thinking. Handle statements with care. Top Next IssueDo you spend your days scurrying from meeting to meeting? Do you ever wonder if all these meetings are really necessary? (They aren't) Or whether there isn't some better way to get this work done? (There is) Read 101 Tips for Effective Meetings to learn how to make meetings much more productive and less stressful — and a lot more rare. Order Now!
Footnotes
Your comments are welcome
Would you like to see your comments posted here? rbrenyrWpTxHuyCrjZbUpner@ChacnoFNuSyWlVzCaGfooCanyon.comSend me your comments by email, or by Web form.About Point Lookout
Thank you for reading this article. I hope you enjoyed it and found it useful, and that you'll consider recommending it to a friend.
This article in its entirety was written by a human being. No machine intelligence was involved in any way.
Point Lookout is a free weekly email newsletter. Browse the archive of past issues. Subscribe for free.
Support Point Lookout by joining the Friends of Point Lookout, as an individual or as an organization.
Do you face a complex interpersonal situation? Send it in, anonymously if you like, and I'll give you my two cents.
Related articles
More articles on Cognitive Biases at Work:
- The Trap of Beautiful Language
- As we assess the validity of others' statements, we risk making a characteristically human error —
we confuse the beauty of their language with the reliability of its meaning. We're easily thrown off
by alliteration, anaphora, epistrophe, and chiasmus.
- Be Choosier About Job Offers: I
- A serious error some job seekers make is accepting an offer that isn't actually a good fit. We make
this mistake for a variety of reasons, including hating the job-search process, desperation, and wishful
thinking. How can we avoid the error?
- The Risk of Astonishing Success
- When we experience success, we're more likely to develop overconfidence. And when the success is so
extreme as to induce astonishment, we become even more vulnerable to overconfidence. It's a real risk
of success that must be managed.
- Lessons Not Learned: I
- The planning fallacy is a cognitive bias that causes us to underestimate the cost and effort involved
in projects large and small. Mitigating its effects requires understanding how we go wrong when we plan
projects by referencing our own past experience.
- Additive Bias…or Not: I
- When we alter existing systems to enhance them, we tend to favor adding components even when subtracting
might be better. This effect has been attributed to a cognitive bias known as additive bias. But other
forces more important might be afoot.
See also Cognitive Biases at Work and Cognitive Biases at Work for more related articles.
Forthcoming issues of Point Lookout
- Coming December 11: White Water Rafting as a Metaphor for Group Development
- Tuckman's model of small group development, best known as "Forming-Storming-Norming-Performing," applies better to development of some groups than to others. We can use a metaphor to explore how the model applies to Storming in task-oriented work groups. Available here and by RSS on December 11.
- And on December 18: Subgrouping and Conway's Law
- When task-oriented work groups address complex tasks, they might form subgroups to address subtasks. The structure of the subgroups and the order in which they form depend on the structure of the group's task and the sequencing of the subtasks. Available here and by RSS on December 18.
Coaching services
I offer email and telephone coaching at both corporate and individual rates. Contact Rick for details at rbrenyrWpTxHuyCrjZbUpner@ChacnoFNuSyWlVzCaGfooCanyon.com or (650) 787-6475, or toll-free in the continental US at (866) 378-5470.
Get the ebook!
Past issues of Point Lookout are available in six ebooks:
- Get 2001-2 in Geese Don't Land on Twigs (PDF, )
- Get 2003-4 in Why Dogs Wag (PDF, )
- Get 2005-6 in Loopy Things We Do (PDF, )
- Get 2007-8 in Things We Believe That Maybe Aren't So True (PDF, )
- Get 2009-10 in The Questions Not Asked (PDF, )
- Get all of the first twelve years (2001-2012) in The Collected Issues of Point Lookout (PDF, )
Are you a writer, editor or publisher on deadline? Are you looking for an article that will get people talking and get compliments flying your way? You can have 500-1000 words in your inbox in one hour. License any article from this Web site. More info
Follow Rick
Recommend this issue to a friend
Send an email message to a friend
rbrenyrWpTxHuyCrjZbUpner@ChacnoFNuSyWlVzCaGfooCanyon.comSend a message to Rick
A Tip A Day feed
Point Lookout weekly feed