Point Lookout: a free weekly publication of Chaco Canyon Consulting
Volume 21, Issue 8;   February 24, 2021: Risk Acceptance: Naïve Realism

Risk Acceptance: Naïve Realism

by

When we suddenly notice a "project-killer" risk that hasn't yet materialized, we sometimes accept the risk even though we know how seriously it threatens the effort. A psychological phenomenon known as naïve realism plays a role in this behavior.
Roger Boisjoly of Morton Thiokol, who tried to halt the launch of the Challenger space shuttle in 1986

Roger Boisjoly, the Morton Thiokol engineer who, in 1985, one year before the catastrophic failure of the U.S. Space Shuttle Challenger, wrote a memorandum outlining the safety risks of cold-weather launches. He successfully raised the issue then, and many times subsequently, including the evening prior to the launch. In 1988, he was awarded the Prize for Scientific Freedom and Responsibility by the American Association for the Advancement of Science, for "…his exemplary and repeated efforts to fulfill his professional responsibilities as an engineer by alerting others to life-threatening design problems of the Challenger space shuttle and for steadfastly recommending against the tragic launch of January 1986."

Mr. Boisjoly did not succeed in gaining acceptance of his objections to the Challenger launch. Moreover, he was shunned by colleagues, even after the investigation concluded, showing that his objections had been correct. He eventually resigned from Morton Thiokol. It is difficult to say from this distance, but this outcome could be an example of excoriation and ejection of a dissenter as a consequence of naïve realism.

Photo courtesy the Online Ethics Center at the National Academy of Engineering.

When a project is underway, or even when we're still in discussion stages, we sometimes stumble upon an unanticipated risk. Some unanticipated risks can be serious — fatal or near fatal to the project if they materialize. For convenience, I call these risks Very Bad News (VBN) risks. After serious study, if we can't devise practical approaches to avoid, transfer, reduce, or compensate for the VBN risk, we cancel the project and move on. That's what happens if we're thinking clearly.

Sometimes we don't think clearly enough. We decide to accept the VBN risk. Here are three scenarios that frequently lead to risk acceptance:

  • Someone who's politically very powerful insists that this project must succeed no matter what, and "you'll find a way around the problem somehow."
  • The team has found some approaches that might mitigate the VBN risk somewhat. The time, budget, or capabilities required are beyond what's available, but somebody says, "If we make a solid demo, they'll find the money."
  • There's division of opinion about whether the VBN risk is real. Most expert members of the project team acknowledge the risk, but one ambitious fellow, currying favor with Management, disagrees.

These situations are fertile ground for toxic conflict. When the VBN risk is fundamentally technological, a common pattern for this conflict is an oppositional debate between the "technologists" (IT or product engineers) on the one hand, and on the other hand the "business" (product managers, executives, or Sales or Marketing officials). When the conflict becomes toxic enough, it can leave lasting social scars that limit future organizational performance.

The conflict that arises in the context of debates about VBN risks isn't unique. We can understand its dynamics in terms of a psychological phenomenon known as naïve realism. Naïve realism is our human tendency to assume that our perceptions of the world are accurate and objective. [Ross 1995] Consequently, when our understanding of the world conflicts with that of others, we attribute the difference to distortions in others' perceptions, due to their ignorance, false beliefs, irrationality, or biases. As we do this, so do our partners in conflict attribute to us ignorance, false beliefs, irrationality, or biases. Naïve realism thus provides an elegantly symmetric setup for toxic conflict. It can transform a disagreement from one in which each party critiques the value of the other's approach to the issues to one in which each party attacks the other's value as a person.

But although a setup for toxic conflict is necessary, it isn't sufficient to seriously damage relationships. For truly destructive toxic conflict, the participants need to care deeply about the outcome of the conflict. VBN risks can provide the missing element. We know that conflict participants care deeply about the outcomes of debates about VBN risks, because they all regard VBN risks as "Very Bad News."

On the "business" side, After we identify a "Very Bad News" risk,
a common pattern for the ensuing
conflict is oppositional debate between
product engineers and the "business"
conflict participants strongly desire project success. They want what the project promises to deliver, and they can get what they want only if the project goes forward. On the "technologist" side, conflict participants want to work on successful projects, and they want to avoid working on projects that are doomed from the start. Too often, in their experience, project failures have been unjustly and incorrectly attributed not to foolhardy decisions to accept VBN risks, but to the lack of professionalism and low work quality of project teams. In the context of the VBN risk, the technologists can get what they want only if the VBN risk is properly acknowledged and managed, the cost of which can be prohibitive. The "business" faction wants the project to go forward despite the VBN risk; the "technologist" faction wants the project to be reconfigured or cancelled because of the VBN risk.

In many organizations, the "business" prevails. In practice, the technologists are directed to execute the project with the VBN risk wholly or largely unmitigated. Because the people who represent the "business" want the project to proceed, and because they cannot afford to manage the VBN risk, they decide to opt for "risk acceptance."

But exactly how does this happen? What patterns of thought and decision making enable the group to proceed despite this stark disagreement? I'll explore this part of the story next time.  Next in this series Go to top Top  Next issue: Risk Acceptance: One Path  Next Issue

How to Spot a Troubled Project Before the Trouble StartsProjects never go quite as planned. We expect that, but we don't expect disaster. How can we get better at spotting disaster when there's still time to prevent it? How to Spot a Troubled Project Before the Trouble Starts is filled with tips for executives, senior managers, managers of project managers, and sponsors of projects in project-oriented organizations. It helps readers learn the subtle cues that indicate that a project is at risk for wreckage in time to do something about it. It's an ebook, but it's about 15% larger than "Who Moved My Cheese?" Just . Order Now! .

Footnotes

Comprehensive list of all citations from all editions of Point Lookout
[Ross 1995]
Lee Ross and Andrew Ward. "Naive realism in everyday life: Implications for social conflict and misunderstanding." (1995). Available here. Back

Your comments are welcome

Would you like to see your comments posted here? rbrendPtoGuFOkTSMQOzxner@ChacEgGqaylUnkmwIkkwoCanyon.comSend me your comments by email, or by Web form.

About Point Lookout

This article in its entirety was written by a 
          human being. No machine intelligence was involved in any way.Thank you for reading this article. I hope you enjoyed it and found it useful, and that you'll consider recommending it to a friend.

This article in its entirety was written by a human being. No machine intelligence was involved in any way.

Point Lookout is a free weekly email newsletter. Browse the archive of past issues. Subscribe for free.

Support Point Lookout by joining the Friends of Point Lookout, as an individual or as an organization.

Do you face a complex interpersonal situation? Send it in, anonymously if you like, and I'll give you my two cents.

Related articles

More articles on Cognitive Biases at Work:

The Great Wall of China near MutianyuScope Creep and Confirmation Bias
As we've seen, some cognitive biases can contribute to the incidence of scope creep in projects and other efforts. Confirmation bias, which causes us to prefer evidence that bolsters our preconceptions, is one of these.
The Leonard P. Zakim Bunker Hill BridgeSeven Planning Pitfalls: III
We usually attribute departures from plan to poor execution, or to "poor planning." But one cause of plan ineffectiveness is the way we think when we set about devising plans. Three cognitive biases that can play roles are the so-called Magical Number 7, the Ambiguity Effect, and the Planning Fallacy.
Auklet flock, Shumagins, March 2006Seven More Planning Pitfalls: I
Planners and members of planning teams are susceptible to patterns of thinking that lead to unworkable plans. But planning teams also suffer vulnerabilities. Two of these are Group Polarization and Trips to Abilene.
Children playing a computer gameThe Risk of Astonishing Success
When we experience success, we're more likely to develop overconfidence. And when the success is so extreme as to induce astonishment, we become even more vulnerable to overconfidence. It's a real risk of success that must be managed.
Roger Boisjoly of Morton Thiokol, who tried to halt the launch of Challenger in 1986The Illusion of Explanatory Depth
The illusion of explanatory depth is the tendency of humans to believe they understand something better than they actually do. Discovering the illusion when you're explaining something is worse than embarrassing. It can be career ending.

See also Cognitive Biases at Work and Conflict Management for more related articles.

Forthcoming issues of Point Lookout

A dangerous curve in an icy roadComing May 1: Antipatterns for Time-Constrained Communication: 2
Recognizing just a few patterns that can lead to miscommunication can reduce the incidence of miscommunications. Here's Part 2 of a collection of antipatterns that arise in communication under time pressure, emphasizing those that depend on content. Available here and by RSS on May 1.
And on May 8: Antipatterns for Time-Constrained Communication: 3
Recognizing just a few patterns that can lead to miscommunication can reduce the incidence of problems. Here is Part 3 of a collection of antipatterns that arise in technical communication under time pressure, emphasizing past experiences of participants. Available here and by RSS on May 8.

Coaching services

I offer email and telephone coaching at both corporate and individual rates. Contact Rick for details at rbrendPtoGuFOkTSMQOzxner@ChacEgGqaylUnkmwIkkwoCanyon.com or (650) 787-6475, or toll-free in the continental US at (866) 378-5470.

Get the ebook!

Past issues of Point Lookout are available in six ebooks:

Reprinting this article

Are you a writer, editor or publisher on deadline? Are you looking for an article that will get people talking and get compliments flying your way? You can have 500-1000 words in your inbox in one hour. License any article from this Web site. More info

Follow Rick

Send email or subscribe to one of my newsletters Follow me at LinkedIn Follow me at X, or share a post Subscribe to RSS feeds Subscribe to RSS feeds
The message of Point Lookout is unique. Help get the message out. Please donate to help keep Point Lookout available for free to everyone.
Technical Debt for Policymakers BlogMy blog, Technical Debt for Policymakers, offers resources, insights, and conversations of interest to policymakers who are concerned with managing technical debt within their organizations. Get the millstone of technical debt off the neck of your organization!
Go For It: Sometimes It's Easier If You RunBad boss, long commute, troubling ethical questions, hateful colleague? Learn what we can do when we love the work but not the job.
303 Tips for Virtual and Global TeamsLearn how to make your virtual global team sing.
101 Tips for Managing ChangeAre you managing a change effort that faces rampant cynicism, passive non-cooperation, or maybe even outright revolt?
101 Tips for Effective MeetingsLearn how to make meetings more productive — and more rare.
Exchange your "personal trade secrets" — the tips, tricks and techniques that make you an ace — with other aces, anonymously. Visit the Library of Personal Trade Secrets.
If your teams don't yet consistently achieve state-of-the-art teamwork, check out this catalog. Help is just a few clicks/taps away!
Ebooks, booklets and tip books on project management, conflict, writing email, effective meetings and more.