
A curious baby. According to Prof. Laura E. Schulz, associate professor of cognitive science in the Brain and Cognitive Sciences department at MIT, babies' approach to learning about the world has much in common with the scientific method. In some sense, we would do better at debugging complex systems if we took a more childlike approach. Photo (cc) Creative Commons Attribution 2.0 by Kenny Louie courtesy Wikimedia.
To repair complex systems, many resort to "random twiddling and part replacement" (RTAPR) when they're under time and resource constraints. Sadly, RTAPR doesn't work very well. For example, consider a system that has six commercial off-the-shelf components. Let's suppose that it isn't working right. We decide to replace Module 2, which produces no change — the system continues to misbehave. Some might conclude that this proves that Module 2 is OK, but that conclusion might be mistaken. Suppose that the problem lies in the firmware of Module 2, which controls how it operates on the data it receives from Module 1. Since both of our Module 2 boxes contained the same firmware, the system behavior didn't change when we made the swap. A conclusion that Module 2 was not involved in the fault would therefore be incorrect.
A more careful approach can work better than RTAPR. Here are some guidelines that form the basis of what is usually called the scientific method.
- Perform no random experiments
- Random experiments, especially those involving system configuration changes, are unlikely to produce new knowledge. The more complicated the system, the less productive are random experiments.
- Keep excellent records
- Record the Random experiments, especially if
they involve system configuration
changes, are unlikely to
produce new knowledgedetails of all experiments and results. Typically, you won't refer to these notes until you're completely stumped, but that happens with alarming frequency for complex systems. So write the notes so as to make them clear in that kind of desperate situation. - Try to replicate unwanted behavior
- (a) If the unwanted behavior is reliably repeatable, observe the results of making a minimal change to the system. Any change in behavior can be revealing. (b) If the unwanted behavior isn't repeatable, try to find a system configuration that makes it repeatable, and then go to (a). In all such experiments, controlling the system's containing environment is essential.
- Base all attempts on hypotheses
- Because the input configuration for a complicated system is also complicated, proving that complicated systems work for all required inputs is difficult. Hypotheses about why the system isn't working are equally difficult to prove. Hypotheses can more readily be disproven than proven.
- Therefore, have a testable hypothesis in mind whenever you change the system configuration. Testable hypotheses are of this form (for example): "The fault might be A. If experiment B produces behavior C, then the fault cannot be A." Repeating this process gradually eliminates possibilities until only the truth remains.
- Fail forward
- Devise hypotheses and experiments that cause your investigation to "fail forward." That is, favor experiments that produce useful knowledge whatever the outcome of the experiment. If you make a change and the system starts working, that should help explain what was wrong. And if that same change causes some other result, that, too, should be enlightening information.
Adhering to these guidelines can be difficult, especially under pressure. If deviation is required, make note of it, and note how deviations affect your conclusions. First issue in this series
Top
Next Issue
Are your projects always (or almost always) late and over budget? Are your project teams plagued by turnover, burnout, and high defect rates? Turn your culture around. Read 52 Tips for Leaders of Project-Oriented Organizations, filled with tips and techniques for organizational leaders. Order Now!
Your comments are welcome
Would you like to see your comments posted here? rbrenDJpmhgyaDTwBQXkhner@ChacmGoYuzfZpOvDQdRkoCanyon.comSend me your comments by email, or by Web form.About Point Lookout
Thank you for reading this article. I hope you enjoyed it and
found it useful, and that you'll consider recommending it to a friend.
This article in its entirety was written by a human being. No machine intelligence was involved in any way.
Point Lookout is a free weekly email newsletter. Browse the archive of past issues. Subscribe for free.
Support Point Lookout by joining the Friends of Point Lookout, as an individual or as an organization.
Do you face a complex interpersonal situation? Send it in, anonymously if you like, and I'll give you my two cents.
Related articles
More articles on Problem Solving and Creativity:
Finger Puzzles and "Common Sense"
- Working on complex projects, we often face a choice between "just do it" and "wait, let's
think this through first." Choosing to just do it can seem to be the shortest path to the goal,
but it rarely is. It's an example of a Finger Puzzle.
Bois Sec!
- When your current approach isn't working, you can scrap whatever you're doing and start again —
if you have enough time and money. There's a less radical solution, and if it works, it's usually both
cheaper and faster.
Managing Wishful Thinking Risk
- When things go wrong, and we look back at how we got there, we must sometimes admit to wishful thinking.
Here's a framework for managing the risk of wishful thinking.
Virtual Brainstorming: I
- When we need to brainstorm, meeting virtually carries a risk that our results might be problematic.
Here's Part I of some steps to take to reduce the risk.
Brainstorming and Speedstorming: II
- Recent research into the effectiveness of brainstorming has raised some questions. Motivated to examine
alternatives, I ran into speedstorming. Here's Part II of an exploration of the properties
of speedstorming.
See also Problem Solving and Creativity and Problem Solving and Creativity for more related articles.
Forthcoming issues of Point Lookout
Coming May 14: Working with the Overconfident
- A cognitive bias known as the Overconfidence Effect causes us to overestimate the reliability of our judgments. Decisions we make based on those judgments are therefore suspect. But there are steps we can take to make our confidence levels more realistic, and thus make our decisions more reliable. Available here and by RSS on May 14.
And on May 21: Mismanaging Project Managers
- Most organizations hold project managers accountable for project performance. But they don't grant those project managers control of needed resources. Nor do they hold project sponsors or other senior managers accountable for the consequences of their actions when they interfere with project work. Here's a catalog of behaviors worth looking at. Available here and by RSS on May 21.
Coaching services
I offer email and telephone coaching at both corporate and individual rates. Contact Rick for details at rbrenDJpmhgyaDTwBQXkhner@ChacmGoYuzfZpOvDQdRkoCanyon.com or (650) 787-6475, or toll-free in the continental US at (866) 378-5470.
Get the ebook!
Past issues of Point Lookout are available in six ebooks:
- Get 2001-2 in Geese Don't Land on Twigs (PDF, )
- Get 2003-4 in Why Dogs Wag (PDF, )
- Get 2005-6 in Loopy Things We Do (PDF, )
- Get 2007-8 in Things We Believe That Maybe Aren't So True (PDF, )
- Get 2009-10 in The Questions Not Asked (PDF, )
- Get all of the first twelve years (2001-2012) in The Collected Issues of Point Lookout (PDF, )
Are you a writer, editor or publisher on deadline? Are you looking for an article that will get people talking and get compliments flying your way? You can have 500-1000 words in your inbox in one hour. License any article from this Web site. More info
Follow Rick
Recommend this issue to a friend
Send an email message to a friend
rbrenDJpmhgyaDTwBQXkhner@ChacmGoYuzfZpOvDQdRkoCanyon.comSend a message to Rick
A Tip A Day feed
Point Lookout weekly feed
