SlideShare a Scribd company logo
1 of 25
Neil Olonoff
 Examples of catastrophic knowledge collapse
 What (really) is COOP?
 Disaster versus “business as usual”
 Building resilience into “business as usual”
 Combine Risk Management and KM
 Knowledge Enabling of Processes
 Identifying “weak signals” in a noisy environment.
1. All were avoidable.
2. All are examples of
catastrophic knowledge
collapse.
Want proof?
Actual telegram containing the observed
position of the iceberg that sank the Titanic.
Documented example of O-ring
blowback on Space Shuttle
Booster Rocket
Cover page of the President’s Daily
Brief, dated August 6,2001, entitled
Bin Ladin Determined to Strike in
US.
Documents Show Early Worries About
Safety of Rig
In each case, a knowledge
gap resulted in a
catastrophic cascade; a
“knowledge collapse”
How?
Small gaps cascade to produce
big impacts. These can be
performance or knowledge gaps
For want of a nail the shoe was lost.
For want of a shoe the horse was lost.
For want of a horse the rider was lost.
For want of a rider the battle was lost.
For want of a battle the kingdom was lost.
And all for the want of a horseshoe.
 “Process breakdown due to knowledge
resource failure.”
 Today, organizations “run on” knowledge
 Knowledge is as essential as gas and oil in a car;
food and water to an organism
 In other words, the flow of knowledge is essential
to operation; stopping the flow causes problems;
if enough problems occur, processes break down.
 Complacency breeds neglect, and ..
 Failure to ”detect weak signals”
 Acceptance of faulty assumptions
 Failure to implement risk management
 Disconnects betweenCOOP and Business as
Usual
 Traditionally “disaster oriented.”
 Views its role only in situations of catastrophe
 Usually provides alternative sites and support
 Usually ignores “business as usual.”
 The difference between “business as usual” and a
disaster is often a matter of degree in terms of the
“number of processes impacted.”
© Neil Olonoff
 Decision and performance support systems
that are integrated with knowledge and
information sources
 Combine KM & Risk Management
 Learn from Failure
 ComprehensiveWho KnowsWhatWhere
resource
 Active, dynamic social networking systems
 Disasters teach more than successes.
 “Never let a good crisis go to waste.”
 Allow learning by exposing failure to the light
of day.
 Knowledge management can help by ..
 Identifying “weak signals”
 Institutionalizing adaptive habits
 Building knowledge into critical processes
 Broad,William J. Taking Lessons FromWhatWentWrong. NewYorkTimes, July 19, 2010.
 Carl, JosephW. Lt Col USAF (Ret) and Freeman,George,Col, USAFR (Ret), Nonstationary Root
Causes of Cobb’s Paradox, DefenseAcquisition Review Journal,Vol 17. No. 55, Jul 2010, pp 337-
1. User Involvement
2. Executive Management Support
3. Clear Statement of Requirements
4. Proper Planning
5. Realistic Expectations
6. Smaller Project Milestones
7. Competent Staff
8. Ownership
9. ClearVision & Objectives
10. Hardworking, Focused Staff
In 1995, Martin Cobb worked for the Secretariat of the
Treasury Board of Canada. He attendedThe Standish
Group’s CHAOS University, where the year’s 10 most
complex information technology (IT) projects are
analyzed and discussed.The 10 most complex IT
projects studied byThe Standish Group in 1994 were
all in trouble: eight were over schedule, on average by
a factor of 1.6 and over budget by a factor of 1.9; the
other two were cancelled and never delivered
anything.That led Cobb to state his now famous
paradox (Cobb, 1995): “We know why [programs] fail;
we know how to prevent their failure—so why do they
still fail?”

More Related Content

Similar to OLONOFF KM FOR COOP 3AUG10

An Evolution in Preparedness
An Evolution in PreparednessAn Evolution in Preparedness
An Evolution in PreparednessDoug Hanchard
 
Transitions Essays
Transitions EssaysTransitions Essays
Transitions EssaysReina Rosado
 
The Key to Great Teams: Understanding the Human Operating System
The Key to Great Teams: Understanding the Human Operating SystemThe Key to Great Teams: Understanding the Human Operating System
The Key to Great Teams: Understanding the Human Operating SystemAtlassian
 
Derailed Crisis Plans
Derailed Crisis PlansDerailed Crisis Plans
Derailed Crisis Plansmschannon
 
HBR.ORG APRIL !## REPRINT R!!#GFAILURE LEARN FROM IT
HBR.ORG APRIL !## REPRINT R!!#GFAILURE LEARN FROM ITHBR.ORG APRIL !## REPRINT R!!#GFAILURE LEARN FROM IT
HBR.ORG APRIL !## REPRINT R!!#GFAILURE LEARN FROM ITJeanmarieColbert3
 
Safety Management Systems (SMS) and Decision Making
Safety Management Systems (SMS) and Decision MakingSafety Management Systems (SMS) and Decision Making
Safety Management Systems (SMS) and Decision MakingIHSTFAA
 
Business Continuity Planning
Business Continuity PlanningBusiness Continuity Planning
Business Continuity PlanningABramall
 
1. What are the differences between the individual rights perspect.docx
1. What are the differences between the individual rights perspect.docx1. What are the differences between the individual rights perspect.docx
1. What are the differences between the individual rights perspect.docxjackiewalcutt
 
US Office of Personnel Management: Notes on "Big Data"
US Office of Personnel Management: Notes on  "Big Data" US Office of Personnel Management: Notes on  "Big Data"
US Office of Personnel Management: Notes on "Big Data" Tom Moritz
 
2.7 workshop-on-intelligence-Steele on future
2.7 workshop-on-intelligence-Steele on future2.7 workshop-on-intelligence-Steele on future
2.7 workshop-on-intelligence-Steele on futureRobert David Steele Vivas
 
Leadership and innovation presentation to UiO Green IT School
Leadership and innovation presentation to UiO Green IT SchoolLeadership and innovation presentation to UiO Green IT School
Leadership and innovation presentation to UiO Green IT SchoolRick Wheatley
 
Lampiran profiting from uncertainty
Lampiran profiting from uncertaintyLampiran profiting from uncertainty
Lampiran profiting from uncertaintyFuturum2
 
Current Event Essay.pdf
Current Event Essay.pdfCurrent Event Essay.pdf
Current Event Essay.pdfLilian Gerlin
 
Cyber Security: The Strategic View
Cyber Security: The Strategic ViewCyber Security: The Strategic View
Cyber Security: The Strategic ViewCisco Canada
 

Similar to OLONOFF KM FOR COOP 3AUG10 (20)

An Evolution in Preparedness
An Evolution in PreparednessAn Evolution in Preparedness
An Evolution in Preparedness
 
Transitions Essays
Transitions EssaysTransitions Essays
Transitions Essays
 
The Key to Great Teams: Understanding the Human Operating System
The Key to Great Teams: Understanding the Human Operating SystemThe Key to Great Teams: Understanding the Human Operating System
The Key to Great Teams: Understanding the Human Operating System
 
Derailed Crisis Plans
Derailed Crisis PlansDerailed Crisis Plans
Derailed Crisis Plans
 
HBR.ORG APRIL !## REPRINT R!!#GFAILURE LEARN FROM IT
HBR.ORG APRIL !## REPRINT R!!#GFAILURE LEARN FROM ITHBR.ORG APRIL !## REPRINT R!!#GFAILURE LEARN FROM IT
HBR.ORG APRIL !## REPRINT R!!#GFAILURE LEARN FROM IT
 
Strong Angel - an Evolution in Preparedness
Strong Angel - an Evolution in PreparednessStrong Angel - an Evolution in Preparedness
Strong Angel - an Evolution in Preparedness
 
Safety Management Systems (SMS) and Decision Making
Safety Management Systems (SMS) and Decision MakingSafety Management Systems (SMS) and Decision Making
Safety Management Systems (SMS) and Decision Making
 
Business Continuity Planning
Business Continuity PlanningBusiness Continuity Planning
Business Continuity Planning
 
Security risk
Security riskSecurity risk
Security risk
 
1. What are the differences between the individual rights perspect.docx
1. What are the differences between the individual rights perspect.docx1. What are the differences between the individual rights perspect.docx
1. What are the differences between the individual rights perspect.docx
 
US Office of Personnel Management: Notes on "Big Data"
US Office of Personnel Management: Notes on  "Big Data" US Office of Personnel Management: Notes on  "Big Data"
US Office of Personnel Management: Notes on "Big Data"
 
2.7 workshop-on-intelligence-Steele on future
2.7 workshop-on-intelligence-Steele on future2.7 workshop-on-intelligence-Steele on future
2.7 workshop-on-intelligence-Steele on future
 
Six Sigma
Six SigmaSix Sigma
Six Sigma
 
Leadership and innovation presentation to UiO Green IT School
Leadership and innovation presentation to UiO Green IT SchoolLeadership and innovation presentation to UiO Green IT School
Leadership and innovation presentation to UiO Green IT School
 
Lampiran profiting from uncertainty
Lampiran profiting from uncertaintyLampiran profiting from uncertainty
Lampiran profiting from uncertainty
 
Current Event Essay.pdf
Current Event Essay.pdfCurrent Event Essay.pdf
Current Event Essay.pdf
 
Current Event Essay
Current Event EssayCurrent Event Essay
Current Event Essay
 
Lecture 5.2CD.pptx
Lecture 5.2CD.pptxLecture 5.2CD.pptx
Lecture 5.2CD.pptx
 
MRAP Program
MRAP ProgramMRAP Program
MRAP Program
 
Cyber Security: The Strategic View
Cyber Security: The Strategic ViewCyber Security: The Strategic View
Cyber Security: The Strategic View
 

OLONOFF KM FOR COOP 3AUG10

  • 2.  Examples of catastrophic knowledge collapse  What (really) is COOP?  Disaster versus “business as usual”  Building resilience into “business as usual”  Combine Risk Management and KM  Knowledge Enabling of Processes  Identifying “weak signals” in a noisy environment.
  • 3.
  • 4.
  • 5.
  • 6.
  • 7. 1. All were avoidable. 2. All are examples of catastrophic knowledge collapse. Want proof?
  • 8. Actual telegram containing the observed position of the iceberg that sank the Titanic.
  • 9. Documented example of O-ring blowback on Space Shuttle Booster Rocket
  • 10. Cover page of the President’s Daily Brief, dated August 6,2001, entitled Bin Ladin Determined to Strike in US.
  • 11. Documents Show Early Worries About Safety of Rig
  • 12. In each case, a knowledge gap resulted in a catastrophic cascade; a “knowledge collapse” How?
  • 13. Small gaps cascade to produce big impacts. These can be performance or knowledge gaps For want of a nail the shoe was lost. For want of a shoe the horse was lost. For want of a horse the rider was lost. For want of a rider the battle was lost. For want of a battle the kingdom was lost. And all for the want of a horseshoe.
  • 14.  “Process breakdown due to knowledge resource failure.”  Today, organizations “run on” knowledge  Knowledge is as essential as gas and oil in a car; food and water to an organism  In other words, the flow of knowledge is essential to operation; stopping the flow causes problems; if enough problems occur, processes break down.
  • 15.  Complacency breeds neglect, and ..  Failure to ”detect weak signals”  Acceptance of faulty assumptions  Failure to implement risk management  Disconnects betweenCOOP and Business as Usual
  • 16.  Traditionally “disaster oriented.”  Views its role only in situations of catastrophe  Usually provides alternative sites and support  Usually ignores “business as usual.”
  • 17.  The difference between “business as usual” and a disaster is often a matter of degree in terms of the “number of processes impacted.” © Neil Olonoff
  • 18.  Decision and performance support systems that are integrated with knowledge and information sources  Combine KM & Risk Management  Learn from Failure  ComprehensiveWho KnowsWhatWhere resource  Active, dynamic social networking systems
  • 19.  Disasters teach more than successes.  “Never let a good crisis go to waste.”  Allow learning by exposing failure to the light of day.  Knowledge management can help by ..  Identifying “weak signals”  Institutionalizing adaptive habits  Building knowledge into critical processes
  • 20.
  • 21.  Broad,William J. Taking Lessons FromWhatWentWrong. NewYorkTimes, July 19, 2010.  Carl, JosephW. Lt Col USAF (Ret) and Freeman,George,Col, USAFR (Ret), Nonstationary Root Causes of Cobb’s Paradox, DefenseAcquisition Review Journal,Vol 17. No. 55, Jul 2010, pp 337-
  • 22.
  • 23. 1. User Involvement 2. Executive Management Support 3. Clear Statement of Requirements 4. Proper Planning 5. Realistic Expectations 6. Smaller Project Milestones 7. Competent Staff 8. Ownership 9. ClearVision & Objectives 10. Hardworking, Focused Staff
  • 24.
  • 25. In 1995, Martin Cobb worked for the Secretariat of the Treasury Board of Canada. He attendedThe Standish Group’s CHAOS University, where the year’s 10 most complex information technology (IT) projects are analyzed and discussed.The 10 most complex IT projects studied byThe Standish Group in 1994 were all in trouble: eight were over schedule, on average by a factor of 1.6 and over budget by a factor of 1.9; the other two were cancelled and never delivered anything.That led Cobb to state his now famous paradox (Cobb, 1995): “We know why [programs] fail; we know how to prevent their failure—so why do they still fail?”

Editor's Notes

  1. My name is Neil Olonoff, and I am the Knowledge Management Lead for HQDA G-4 (Logistics) in the Pentagon. I’m also the Co-Chair of the Federal KM Working Group, we are 700 Federal employees and contractors who are working for KM in the Government.
  2. What is catastrophic knowledge collapse? Small knowledge problems can result in process breakdown What (really) is COOP? Disaster scenarios versus “business as usual.” A continuum. Building resilience into “business as usual” Combine Risk Management and KM Knowledge Enabling of Processes Identifying Weak Signals Weak signals in a noisy environment? I’ve got a couple of dogs, and I like to walk them near a lake near my house. There’s a little creekbed with a bridge and the dogs run in the creekbed. I’m not what you’d call a naturalist, but I do look around to see if there’s anything that might bite me, anyway. Well in the little creekbed, just after a big rainstorm, I saw the oddest little creature. It was orange, with yellow spots. Salamander, I thought. But looking closer I saw it had a spiky little body, kind of fat instead of skinny. And a spiky head and a spiky pointed tail. It wasn’t moving so I grabbed a stick to poke it. It just turned over. The back was cream colored and it didn’t move at all, and that’s when I realized it was a child’s toy plastic dinosaur. Weak signals against a noisy environment. If id seen it on a sidewalk instead of a creek bed I woudn’t have given it a second thought.
  3. On April __ the Titanic sank. It hit an iceberg. How many died? As we all know, resulting in the tragic death of Leonardo di Caprio and hundreds of other non-essential personnel.
  4. The space shuttle challenger blasted off on a very cool morning. The o-rings on the booster rockets gave way, resulting in the explosion that killed the crew.
  5. 9/11. I work in a location just adjacent to the area that was struck. You could say that the reverberations of this event still are being felt, especially in the political and military events that are occurring today in Iraq, Afghanistan and elsewhere.
  6. It seems that this week they’ll finally be able to cap the well after ___ weeks of oil spilling into the gulf of mexico.
  7. I understand that this was just one of several telegrams. But this was the crucial one that, if seen by the captain, would have avoided catastrophe.
  8. The Challenger: NASA / JPL Contractor Morton Thiokol and NASA engineers knew about --- and had documented -– the danger of booster rocket O-ring blowback when launching in cool weathers. But because the President wanted to talk about the “teacher in space” in the State of the Union speech, NASA executives overruled the engineers.
  9. Bin Ladin Determined To Strike in US was the President's Daily Brief given to U.S. President George W. Bush on August 6, 2001. The President's Daily Brief (PDB) is a brief of important classified information on national security collected by various U.S. intelligence agencies given to the president and a select group of senior officials. The brief warned of terrorism threats from Osama bin Laden and al-Qaeda over a month before the September 11, 2001 attacks.[1]
  10. May 29, 2010 Documents Show Early Worries About Safety of Rig By IAN URBINA WASHINGTON — Internal documents from BP show that there were serious problems and safety concerns with the Deepwater Horizon rig far earlier than those the company described to Congress last week. The problems involved the well casing and the blowout preventer, which are considered critical pieces in the chain of events that led to the disaster on the rig. The documents show that in March, after several weeks of problems on the rig, BP was struggling with a loss of “well control.” And as far back as 11 months ago, it was concerned about the well casing and the blowout preventer. On June 22, for example, BP engineers expressed concerns that the metal casing the company wanted to use might collapse under high pressure. “This would certainly be a worst-case scenario,” Mark E. Hafle, a senior drilling engineer at BP, warned in an internal report. “However, I have seen it happen so know it can occur.” The company went ahead with the casing, but only after getting special permission from BP colleagues because it violated the company’s safety policies and design standards. The internal reports do not explain why the company allowed for an exception. BP documents released last week to The Times revealed that company officials knew the casing was the riskier of two options. Though his report indicates that the company was aware of certain risks and that it made the exception, Mr. Hafle, testifying before a panel on Friday in Louisiana about the cause of the rig disaster, rejected the notion that the company had taken risks. “Nobody believed there was going to be a safety issue,” Mr. Hafle told a six-member panel of Coast Guard and Minerals Management Service officials. “All the risks had been addressed, all the concerns had been addressed, and we had a model that suggested if executed properly we would have a successful job,” he said. Mr. Hafle, asked for comment by a reporter after his testimony Friday about the internal report, declined to answer questions. BP’s concerns about the casing did not go away after Mr. Hafle’s 2009 report. In April of this year, BP engineers concluded that the casing was “unlikely to be a successful cement job,” according to a document, referring to how the casing would be sealed to prevent gases from escaping up the well. The document also says that the plan for casing the well is “unable to fulfill M.M.S. regulations,” referring to the Minerals Management Service. A second version of the same document says “It is possible to obtain a successful cement job” and “It is possible to fulfill M.M.S. regulations.” Andrew Gowers, a BP spokesman, said the second document was produced after further testing had been done. On Tuesday Congress released a memorandum with preliminary findings from BP’s internal investigation, which indicated that there were warning signs immediately before the explosion on April 20, including equipment readings suggesting that gas was bubbling into the well, a potential sign of an impending blowout. A parade of witnesses at hearings last week told about bad decisions and cut corners in the days and hours before the explosion of the rig, but BP’s internal documents provide a clearer picture of when company and federal officials saw problems emerging. In addition to focusing on the casing, investigators are also focusing on the blowout preventer, a fail-safe device that was supposed to slice through a drill pipe in a last-ditch effort to close off the well when the disaster struck. The blowout preventer did not work, which is one of the reasons oil has continued to spill into the gulf, though the reason it failed remains unclear. Federal drilling records and well reports obtained through the Freedom of Information Act and BP’s internal documents, including more than 50,000 pages of company e-mail messages, inspection reports, engineering studies and other company records obtained by The Times from Congressional investigators, shed new light on the extent and timing of problems with the blowout preventer and the casing long before the explosion. Kendra Barkoff, a spokeswoman for the Interior Department, declined to answer questions about the casings, the blowout preventer and regulators’ oversight of the rig because those matters are part of a continuing investigation. The documents show that in March, after problems on the rig that included drilling mud falling into the formation, sudden gas releases known as “kicks” and a pipe falling into the well, BP officials informed federal regulators that they were struggling with a loss of “well control.” On at least three occasions, BP records indicate, the blowout preventer was leaking fluid, which the manufacturer of the device has said limits its ability to operate properly. “The most important thing at a time like this is to stop everything and get the operation under control,” said Greg McCormack, director of the Petroleum Extension Service at the University of Texas, Austin, offering his assessment about the documents. He added that he was surprised that regulators and company officials did not commence a review of whether drilling should continue after the well was brought under control. After informing regulators of their struggles, company officials asked for permission to delay their federally mandated test of the blowout preventer, which is supposed to occur every two weeks, until the problems were resolved, BP documents say. At first, the minerals agency declined. “Sorry, we cannot grant a departure on the B.O.P. test further than when you get the well under control,” wrote Frank Patton, a minerals agency official. But BP officials pressed harder, citing “major concerns” about doing the test the next day. And by 10:58 p.m., David Trocquet, another M.M.S. official, acquiesced. “After further consideration,” Mr. Trocquet wrote, “an extension is approved to delay the B.O.P. test until the lower cement plug is set.” When the blowout preventer was eventually tested again, it was tested at a lower pressure — 6,500 pounds per square inch — than the 10,000-pounds-per-square-inch tests used on the device before the delay. It tested at this lower pressure until the explosion. A review of Minerals Management Service’s data of all B.O.P. tests done in deep water in the Gulf of Mexico for five years shows B.O.P. tests rarely dropped so sharply, and, in general, either continued at the same threshold or were done at increasing levels. The manufacturer of the blowout preventer, Cameron, declined to say what the appropriate testing pressure was for the device. In an e-mail message, Mr. Gowers of BP wrote that until their investigation was complete, it was premature to answer questions about the casings or the blowout preventer. Even though the documents asking regulators about testing the blowout preventer are from BP, Mr. Gowers said that any questions regarding the device should be directed to Transocean, which owns the rig and, he said, was responsible for maintenance and testing of the device. Transocean officials declined to comment. Bob Sherrill, an expert on blowout preventers and the owner of Blackwater Subsea, an engineering consulting firm, said the conditions on the rig in February and March and the language used by the operator referring to a loss of well control “sounds like they were facing a blowout scenario.” Mr. Sherrill said federal regulators made the right call in delaying the blowout test, because doing a test before the well is stable risks gas kicks. But once the well was stable, he added, it would have made sense for regulators to investigate the problems further. In April, the month the rig exploded, workers encountered obstructions in the well. Most of the problems were conveyed to federal regulators, according to federal records. Many of the incidents required that BP get a permit for a new tactic for dealing with the problem. One of the final indications of such problems was an April 15 request for a permit to revise its plan to deal with a blockage, according to federal documents obtained from Congress by the Center for Biological Diversity, an environmental advocacy group. In the documents, company officials apologized to federal regulators for not having mentioned the type of casing they were using earlier, adding that they had “inadvertently” failed to include it. In the permit request, they did not disclose BP’s own internal concerns about the design of the casing. Less than 10 minutes after the request was submitted, federal regulators approved the permit. Robbie Brown contributed reporting from Kenner, La., and Andy Lehren from New York.
  11. Small gaps cascade to produce big impacts. These can be performance or knowledge gaps This characterizes the cascading 2nd order and 3rd order effects. For want of a nail the shoe was lost. For want of a shoe the horse was lost. For want of a horse the rider was lost. For want of a rider the battle was lost. For want of a battle the kingdom was lost. And all for the want of a horseshoe.
  12. What Does “Knowledge Collapse” Really Mean? “Process breakdown due to knowledge resource failure.” Today, organizations “run on” knowledge Knowledge is as essential as gas and oil in a car; food and water to an organism In other words, the flow of knowledge is essential to operation; stopping the flow causes problems; if enough problems occur, processes break down.
  13. How “gaps” occur Complacency breeds neglect, and .. Failure to ”detect weak signals” Acceptance of faulty assumptions Failure to implement risk management Disconnects between COOP and Business as Usual
  14. The difference between “business as usual” and a disaster is often a matter of degree in terms of the “number of processes impacted.” In other words, disaster often is different from business as usual in terms of the degree, but not the type of dysfunction. The implications of this are: It is possible to plan or prepare for disaster by building resilience into “Business as usual.” The same processes
  15. Building Resilience into “Business as Usual” Decision and performance support systems that are integrated with knowledge and information sources Combine KM & Risk Management See the article “Fusing Risk Management & Knowledge Management” by .. Knowledge Based Risks (KBR’s) Infuse knowledge into work processes - Close knowledge gaps by: -providing broader access to risk information -capture and transfer what you learn along the way Identify Risks -Lessons Learned – often these were risks that challenged an earlier program KBRs provide that, plus … Analysis and planning info Four Practices 1. Pause and Learn (based on US Army’s After Action Review program) 2. Knowledge Sharing Forums 3. Experience based training using case studies 4. Web-enabled teams Learn from Failure Comprehensive Who Knows What Where resource Active, dynamic social networking systems
  16. Reprints This copy is for your personal, noncommercial use only. You can order presentation-ready copies for distribution to your colleagues, clients or customers here or use the "Reprints" tool that appears next to any article. Visit www.nytreprints.com for samples and additional information. Order a reprint of this article now. July 19, 2010 Taking Lessons From What Went Wrong By WILLIAM J. BROAD Disasters teach more than successes. While that idea may sound paradoxical, it is widely accepted among engineers. They say grim lessons arise because the reasons for triumph in matters of technology are often arbitrary and invisible, whereas the cause of a particular failure can frequently be uncovered, documented and reworked to make improvements. Disaster, in short, can become a spur to innovation. There is no question that the trial-and-error process of building machines and industries has, over the centuries, resulted in the loss of much blood and many thousands of lives. It is not that failure is desirable, or that anyone hopes for or aims for a disaster. But failures, sometimes appalling, are inevitable, and given this fact, engineers say it pays to make good use of them to prevent future mistakes. The result is that the technological feats that define the modern world are sometimes the result of events that some might wish to forget. “It’s a great source of knowledge — and humbling, too — sometimes that’s necessary,” said Henry Petroski, a historian of engineering at Duke University and author of “Success Through Failure,” a 2006 book. “Nobody wants failures. But you also don’t want to let a good crisis go to waste.” Now, experts say, that kind of analysis will probably improve the complex gear and procedures that companies use to drill for oil in increasingly deep waters. They say the catastrophic failure involving the Deepwater Horizon oil rig in the Gulf of Mexico on April 20 — which took 11 lives and started the worst offshore oil spill in United States history — will drive the technological progress. “The industry knows it can’t have that happen again,” said David W. Fowler, a professor at the University of Texas, Austin, who teaches a course on forensic engineering. “It’s going to make sure history doesn’t repeat itself.” One possible lesson of the disaster is the importance of improving blowout preventers — the devices atop wells that cut off gushing oil in emergencies. The preventer on the runaway well failed. Even before the disaster, the operators of many gulf rigs had switched to more advanced preventers, strengthening this last line of defense. Of course, an alternative to improving a particular form of technology might be to discard it altogether as too risky or too damaging. Abandoning offshore drilling is certainly one result that some environmentalists would push for — and not only because of potential disasters like the one in the gulf. They would rather see technologies that pump carbon into the atmosphere, threatening to speed global climate change, go extinct than evolve. In London on June 22 at the World National Oil Companies Congress, protesters from Greenpeace interrupted an official from BP, the company that dug the runaway well. Planetary responsibility, a protestor shouted before being taken away, “means stopping the push for dangerous drilling in deep waters.” The history of technology suggests that such an end is unlikely. Devices fall out of favor, but seldom if ever get abolished by design. The explosion of the Hindenburg showed the dangers of hydrogen as a lifting gas and resulted in new emphasis on helium, which is not flammable, rather than ending the reign of rigid airships. And engineering, by definition, is a problem-solving profession. Technology analysts say that constructive impulse, and its probable result for deep ocean drilling, is that innovation through failure analysis will make the wells safer, whatever the merits of reducing human reliance on oil. They hold that the BP disaster, like countless others, will ultimately inspire technological advance. The sinking of the Titanic, the meltdown of the Chernobyl reactor in 1986, the collapse of the World Trade Center — all forced engineers to address what came to be seen as deadly flaws. “Any engineering failure has a lot of lessons,” said Gary Halada, a professor at the State University of New York at Stony Brook who teaches a course called “Learning from Disaster.” Design engineers say that, too frequently, the nature of their profession is to fly blind. Eric H. Brown, a British engineer who developed aircraft during World War II and afterward taught at Imperial College London, candidly described the predicament. In a 1967 book, he called structural engineering “the art of molding materials we do not really understand into shapes we cannot really analyze, so as to withstand forces we cannot really assess, in such a way that the public does not really suspect.” Among other things, Dr. Brown taught failure analysis. Dr. Petroski, at Duke, writing in “Success Through Failure,” noted the innovative corollary. Failures, he said, “always teach us more than the successes about the design of things. And thus the failures often lead to redesigns — to new, improved things.” One of his favorite examples is the 1940 collapse of the Tacoma Narrows Bridge. The span, at the time the world’s third-longest suspension bridge, crossed a strait of Puget Sound near Tacoma, Wash. A few months after its opening, high winds caused the bridge to fail in a roar of twisted metal and shattered concrete. No one died. The only fatality was a black cocker spaniel named Tubby. Dr. Petroski said the basic problem lay in false confidence. Over the decades, engineers had built increasingly long suspension bridges, with each new design more ambitious. The longest span of the Brooklyn Bridge, which opened to traffic in 1883, was 1,595 feet. The George Washington Bridge (1931) more than doubled that distance to 3,500 feet. And the Golden Gate Bridge (1937) went even farther, stretching its middle span to 4,200 feet. “This is where success leads to failure,” Dr. Petroski said in an interview. “You’ve got all these things working. We want to make them longer and more slender.” The Tacoma bridge not only possessed a very long central span — 2,800 feet — but its concrete roadway consisted of just two lanes and its deck was quite shallow. The wind that day caused the insubstantial thoroughfare to undulate wildly up and down and then disintegrate. (A 16-millimeter movie camera capturedthe violent collapse.) Teams of investigators studied the collapse carefully, and designers of suspension bridges took away several lessons. The main one was to make sure the road’s weight and girth were sufficient to avoid risky perturbations from high winds. Dr. Petroski said the collapse had a direct impact on the design of the Verrazano-Narrows Bridge, which opened in 1964 to link Brooklyn and Staten Island. Its longest span was 4,260 feet — making it, at the time, the world’s longest suspension bridge and potentially a disaster-in-waiting. To defuse the threat of high winds, the designers from the start made the roadway quite stiff and added a second deck, even though the volume of traffic was insufficient at first to warrant the lower one. The lower deck remained closed to traffic for five years, opening in 1969. “Tacoma Narrows changed the way that suspension bridges were built,” Dr. Petroski said. “Before it happened, bridge designers didn’t take the wind seriously.” Another example in learning from disaster centers on an oil drilling rig called Ocean Ranger. In 1982, the rig, the world’s largest, capsized and sank off Newfoundland in a fierce winter storm, killing all 84 crew members. The calamity is detailed in a 2001 book, “Inviting Disaster: Lessons from the Edge of Technology,” by James R. Chiles. The floating rig, longer than a football field and 15 stories high, had eight hollow legs. At the bottom were giant pontoons that crewmen could fill with seawater or pump dry, raising the rig above the largest storm waves — in theory, at least. The night the rig capsized, the sea smashed in a glass porthole in the pontoon control room, soaking its electrical panel. Investigators found that the resulting short circuits began a cascade of failures and miscalculations that resulted in the rig’s sinking. The lessons of the tragedy included remembering to shut watertight storm hatches over glass windows, buying all crew members insulated survival suits (about $450 each at the time) and rethinking aspects of rig architecture. “It was a terrible design,” said Dr. Halada of the State University of New York. “But they learned from it.” Increasingly, such tragedies get studied, and not just at Stony Brook. The Stanford University Center for Professional Development offers a graduate certificate in advanced structures and failure analysis. Drexel University offers a master’s degree in forensic science with a focus on engineering. So too, professional engineering has produced a subspecialty that investigates disasters. One of the biggest names in the business is Exponent, a consulting company based in Menlo Park, Calif. It has a staff of 900 specialists around the globe with training in 90 engineering and scientific fields. Exponent says its analysts deal with everything from cars and roller coasters to oil rigs and hip replacements. “We analyze failures and accidents,” the company says, “to determine their causes and to understand how to prevent them.” Forensic engineers say it is too soon to know what happened with Deepwater Horizon, whose demise flooded the gulf with crude oil. They note that numerous federal agencies are involved in a series of detailed investigations, and that President Obama has appointed a blue-ribbon commission to make recommendations on how to strengthen federal oversight of oil rigs. But the engineers hold, seemingly with one voice, that the investigatory findings will eventually improve the art of drilling for oil in deep waters — at least until the next unexpected tragedy, and the next lesson in making the technology safer. One lesson might be to build blowout preventers with more than one blind shear ram. In an emergency, the massive blades of these devices slice through the drill pipe to cut off the flow of gushing oil. The Deepwater Horizon had just one, while a third of the rigs in the gulf now have two. Perhaps regulators will decided that rig operators, whatever the cost, should install more blind shear rams on all blowout preventers. “It’s like our personal lives,” said Dr. Fowler of the University of Texas. “Failure can force us to make hard decisions.” More in Science (2 of 41 articles) Adventures in Very Recent Evolution Read More » Close
  17. Project Success Criteria used by Standish 1. User Involvement 2. Executive Management Support 3. Clear Statement of Requirements 4. Proper Planning 5. Realistic Expectations 6. Smaller Project Milestones 7. Competent Staff 8. Ownership 9. Clear Vision & Objectives 10. Hardworking, Focused Staff
  18. Cobb’s Paradox In 1995, Martin Cobb worked for the Secretariat of the Treasury Board of Canada. He attended The Standish Group’s CHAOS University, where the year’s 10 most complex information technology (IT) projects are analyzed and discussed. The 10 most complex IT projects studied by The Standish Group in 1994 were all in trouble: eight were over schedule, on average by a factor of 1.6 and over budget by a factor of 1.9; the other two were cancelled and never delivered anything. That led Cobb to state his nowfamous paradox (Cobb, 1995): “We know why [programs] fail; we know how to prevent their failure—so why do they still fail?” Cobb’s Paradox states, “We know why [programs] fail; we know how to prevent their failure—so why do they still fail?” One possibility is that we do not really know why programs fail and there is no paradox. Another possibility is that some of the problems that lead to program failure may not be susceptible to practical solution, so that continued failure is not paradoxical. This article defines what we mean by nonstationary root causes of program failures, and identifies 10 such causes. Requirements volatility, funding stability, process immaturity, and lack of discipline are often cited among the reasons. The article ends with recommended approaches to mitigate the effects of influences from the environment that change over time—nonstationary effects. In 2007, the many examples of government project failures led then- Under Secretary of Defense for Acquisition, Technology and Logistics John Young to issue a memorandum that requires prototyping and competition on all major programs up to Milestone B (Young, 2007). Young’s memorandum was a propitious start. But is it likely to be sufficient to solve all the problems that lead to project failure? This article summarizes the number and spectrum of project failures, and makes the case that project failures cannot be attributed solely to mismanagement on the part of project managers. Rather, it appears improbable that all project managers of large complex projects could produce similar failures. The prevailing perception throughout the acquisition community is that program and project managers know why projects fail and how to prevent them from failing. The authors discuss the concept of other influences from the environment that change over time—nonstationary effects—that may be the root cause of these numerous project failures. Background In 2006, a Government Accountability Office report (GAO, 2006) highlighted several government project failures. In the last 5 years, the Department of Defense (DoD) has doubled its planned investments in new weapon systems from about $700 billion in 2001 to nearly $1.4 trillion in 2006. While the weapons that DoD develops have no rival in superiority, weapon systems acquisition remains a long-standing, high-risk area. GAO's reviews over the past 30 years have found consistent problems with weapon acquisitions such as cost increases, schedule delays, and performance shortfalls. The report goes on to state that this huge increase in spending over the past 5 years “has not been accompanied by more stability, better outcomes, or more buying power for the acquisition dollar.” Examples of this huge increase in spending follow: • Capable satellites, potential overrun of $1.4 billion • Satellite payload cost and schedule overruns greater than $1.1 billion • Radar contract projected to overrun target cost by up to 34 percent • Advanced Precision Kill Weapon System (Joint Attack Munition Systems), curtailment of initial program in January 2005 due to development cost overruns, projected schedule slip of 1–2 years, unsatisfactory contract performance, and environmental issues • C-5 Avionics Modernization Program, $23 million cost overrun • C-5 Reliability Enhancement and Re-engineering Program, $209 million overrun • F-22A, increase in the costs of avionics since 1997 by more than $951 million or 24 percent, and other problems discovered late in the program. On March 31, 2006, Comptroller General of the United States David M. Walker stated in congressional testimony: The cost of developing a weapon system continues to often exceed estimates by approximately 30 percent to 40 percent. This in turn results in fewer quantities, missed deadlines, and performance shortfalls. In short, the buying power of the weapon system investment dollar is reduced, the warfighter gets less than promised, and opportunities to make other investments are lost. This is not to say that the nation does not get superior weapons in the end, but that at twice the level of investment. DoD has an obligation to get better results. In the larger context, DoD needs to make changes…consistent with getting the desired outcomes from the acquisition process. Cobb’s Paradox In 1995, Martin Cobb worked for the Secretariat of the Treasury Board of Canada. He attended The Standish Group’s CHAOS University, where the year’s 10 most complex information technology (IT) projects are analyzed and discussed. The 10 most complex IT projects studied by The Standish Group in 1994 were all in trouble: eight were over schedule, on average by a factor of 1.6 and over budget by a factor of 1.9; the other two were cancelled and never delivered anything. That led Cobb to state his nowfamous paradox (Cobb, 1995): “We know why [programs] fail; we know how to prevent their failure—so why do they still fail?” The Standish Group uses project success criteria from surveyed IT managers to create a success-potential chart. The success criteria are shown in the Table, where they are ranked according to their perceived importance. There seems to be an assumption that all the criteria are stationary—that they are assumed to be present on any specific project to some degree and do not change over time except potentially for the better with conscious effort. A little more formally, a process or system is said to be stationary if its behavioral description does not change over time, and nonstationary if its behavioral description does change over time.