• Tweet

  • Mail service

  • Share

  • Save

  • Get PDF

  • Buy Copies

The wisdom of learning from failure is incontrovertible. Yet organizations that do it well are extraordinarily rare. This gap is not due to a lack of commitment to learning. Managers in the vast majority of enterprises that I have studied over the past 20 years—pharmaceutical, financial services, production blueprint, telecommunication, and construction companies; hospitals; and NASA's space shuttle program, among others—genuinely wanted to help their organizations learn from failures to meliorate future performance. In some cases they and their teams had devoted many hours to afterwards-activeness reviews, postmortems, and the like. But time after time I saw that these painstaking efforts led to no existent modify. The reason: Those managers were thinking near failure the wrong way.

Most executives I've talked to believe that failure is bad (of class!). They besides believe that learning from it is pretty straightforward: Ask people to reflect on what they did wrong and exhort them to avoid similar mistakes in the future—or, meliorate all the same, assign a squad to review and write a written report on what happened and and so distribute information technology throughout the organization.

These widely held beliefs are misguided. Get-go, failure is not always bad. In organizational life it is sometimes bad, sometimes inevitable, and sometimes fifty-fifty skilful. 2d, learning from organizational failures is anything but straightforward. The attitudes and activities required to finer detect and analyze failures are in short supply in virtually companies, and the need for context-specific learning strategies is underappreciated. Organizations demand new and better ways to go beyond lessons that are superficial ("Procedures weren't followed") or self-serving ("The market just wasn't ready for our swell new product"). That means jettisoning sometime cultural beliefs and stereotypical notions of success and embracing failure's lessons. Leaders can begin past understanding how the blame game gets in the way.

The Blame Game

Failure and error are virtually inseparable in almost households, organizations, and cultures. Every child learns at some point that albeit failure means taking the blame. That is why so few organizations accept shifted to a culture of psychological safety in which the rewards of learning from failure tin be fully realized.

Executives I've interviewed in organizations equally dissimilar as hospitals and investment banks admit to being torn: How can they respond constructively to failures without giving ascension to an anything-goes attitude? If people aren't blamed for failures, what volition ensure that they endeavor equally hard every bit possible to practise their all-time work?

This concern is based on a false dichotomy. In actuality, a culture that makes information technology safe to admit and report on failure can—and in some organizational contexts must—coexist with loftier standards for performance. To understand why, expect at the exhibit "A Spectrum of Reasons for Failure," which lists causes ranging from deliberate deviation to thoughtful experimentation.

Which of these causes involve blameworthy actions? Deliberate deviance, start on the listing, obviously warrants blame. But inattention might not. If it results from a lack of effort, perhaps it's blameworthy. But if it results from fatigue well-nigh the terminate of an overly long shift, the manager who assigned the shift is more than at fault than the employee. As nosotros get downwardly the list, it gets more than and more difficult to discover blameworthy acts. In fact, a failure resulting from thoughtful experimentation that generates valuable information may actually exist praiseworthy.

When I ask executives to consider this spectrum and then to estimate how many of the failures in their organizations are truly blameworthy, their answers are usually in single digits—perhaps ii% to 5%. Only when I ask how many are treated equally blameworthy, they say (later a intermission or a laugh) lxx% to xc%. The unfortunate consequence is that many failures go unreported and their lessons are lost.

Not All Failures Are Created Equal

A sophisticated understanding of failure'south causes and contexts will help to avoid the blame game and institute an constructive strategy for learning from failure. Although an space number of things can go wrong in organizations, mistakes autumn into three broad categories: preventable, complication-related, and intelligent.

Preventable failures in predictable operations.

Well-nigh failures in this category can indeed be considered "bad." They unremarkably involve deviations from spec in the closely defined processes of high-book or routine operations in manufacturing and services. With proper preparation and support, employees can follow those processes consistently. When they don't, deviance, inattention, or lack of ability is ordinarily the reason. But in such cases, the causes tin be readily identified and solutions developed. Checklists (equally in the Harvard surgeon Atul Gawande's recent best seller The Checklist Manifesto) are ane solution. Another is the vaunted Toyota Production Arrangement, which builds continual learning from tiny failures (small process deviations) into its approach to improvement. As virtually students of operations know well, a team fellow member on a Toyota assembly line who spots a trouble or fifty-fifty a potential problem is encouraged to pull a rope called the andon string, which immediately initiates a diagnostic and problem-solving process. Production continues unimpeded if the problem can exist remedied in less than a infinitesimal. Otherwise, production is halted—despite the loss of acquirement entailed—until the failure is understood and resolved.

Unavoidable failures in complex systems.

A large number of organizational failures are due to the inherent doubt of work: A detail combination of needs, people, and issues may have never occurred before. Triaging patients in a hospital emergency room, responding to enemy actions on the battleground, and running a fast-growing start-up all occur in unpredictable situations. And in circuitous organizations like shipping carriers and nuclear power plants, system failure is a perpetual run a risk.

Although serious failures can be averted by post-obit best practices for safety and gamble management, including a thorough analysis of whatsoever such events that exercise occur, modest process failures are inevitable. To consider them bad is non just a misunderstanding of how circuitous systems work; it is counterproductive. Avoiding consequential failures means apace identifying and correcting small failures. About accidents in hospitals effect from a series of small failures that went unnoticed and unfortunately lined up in but the wrong fashion.

Intelligent failures at the frontier.

Failures in this category tin rightly be considered "good," considering they provide valuable new knowledge that can help an organization leap ahead of the competition and ensure its future growth—which is why the Duke University professor of management Sim Sitkin calls them intelligent failures. They occur when experimentation is necessary: when answers are not knowable in advance considering this exact situation hasn't been encountered before and perhaps never will be over again. Discovering new drugs, creating a radically new business, designing an innovative product, and testing customer reactions in a brand-new marketplace are tasks that require intelligent failures. "Trial and error" is a common term for the kind of experimentation needed in these settings, but information technology is a misnomer, because "error" implies that there was a "right" outcome in the first identify. At the frontier, the right kind of experimentation produces good failures quickly. Managers who practice it tin avoid the unintelligent failure of conducting experiments at a larger calibration than necessary.

Leaders of the product design firm IDEO understood this when they launched a new innovation-strategy service. Rather than assistance clients pattern new products inside their existing lines—a process IDEO had all only perfected—the service would assistance them create new lines that would accept them in novel strategic directions. Knowing that it hadn't yet figured out how to deliver the service effectively, the company started a minor project with a mattress company and didn't publicly announce the launch of a new concern.

Although the project failed—the customer did non change its product strategy—IDEO learned from information technology and figured out what had to exist done differently. For instance, information technology hired team members with MBAs who could meliorate help clients create new businesses and made some of the clients' managers role of the squad. Today strategic innovation services business relationship for more than a third of IDEO'south revenues.

Tolerating unavoidable process failures in complex systems and intelligent failures at the frontiers of knowledge won't promote mediocrity. Indeed, tolerance is essential for any organization that wishes to extract the knowledge such failures provide. But failure is still inherently emotionally charged; getting an organization to have it takes leadership.

Edifice a Learning Civilisation

Just leaders can create and reinforce a culture that counteracts the blame game and makes people experience both comfortable with and responsible for surfacing and learning from failures. (Meet the sidebar "How Leaders Can Build a Psychologically Safe Environment.") They should insist that their organizations develop a clear understanding of what happened—non of "who did information technology"—when things go wrong. This requires consistently reporting failures, small and big; systematically analyzing them; and proactively searching for opportunities to experiment.

Leaders should also transport the correct bulletin about the nature of the work, such every bit reminding people in R&D, "Nosotros're in the discovery concern, and the faster we fail, the faster nosotros'll succeed." I have found that managers often don't empathize or appreciate this subtle but crucial point. They also may approach failure in a manner that is inappropriate for the context. For example, statistical procedure control, which uses information assay to assess unwarranted variances, is not practiced for catching and correcting random invisible glitches such every bit software bugs. Nor does it help in the development of creative new products. Conversely, though corking scientists intuitively adhere to IDEO'south slogan, "Fail often in order to succeed sooner," it would hardly promote success in a manufacturing establish.

The slogan "Fail often in order to succeed sooner" would inappreciably promote success in a manufacturing constitute.

Often 1 context or 1 kind of work dominates the culture of an enterprise and shapes how information technology treats failure. For instance, automotive companies, with their predictable, loftier-volume operations, understandably tend to view failure as something that tin and should be prevented. But about organizations engage in all three kinds of work discussed above—routine, complex, and borderland. Leaders must ensure that the correct arroyo to learning from failure is applied in each. All organizations learn from failure through three essential activities: detection, analysis, and experimentation.

Detecting Failure

Spotting big, painful, expensive failures is like shooting fish in a barrel. Merely in many organizations any failure that can be subconscious is hidden every bit long equally it'southward unlikely to cause firsthand or obvious damage. The goal should be to surface it early on, before it has mushroomed into disaster.

Presently after arriving from Boeing to accept the reins at Ford, in September 2006, Alan Mulally instituted a new organization for detecting failures. He asked managers to color code their reports green for good, yellow for circumspection, or scarlet for problems—a common management technique. Co-ordinate to a 2009 story in Fortune, at his first few meetings all the managers coded their operations green, to Mulally'due south frustration. Reminding them that the company had lost several billion dollars the previous year, he asked straight out, "Isn't anything not going well?" After ane tentative yellow report was made virtually a serious product defect that would probably delay a launch, Mulally responded to the deathly silence that ensued with applause. After that, the weekly staff meetings were full of color.

That story illustrates a pervasive and primal problem: Although many methods of surfacing current and pending failures be, they are grossly underutilized. Total Quality Management and soliciting feedback from customers are well-known techniques for bringing to light failures in routine operations. Loftier-reliability-organization (HRO) practices help prevent catastrophic failures in complex systems like nuclear power plants through early detection. Electricité de France, which operates 58 nuclear ability plants, has been an exemplar in this area: It goes beyond regulatory requirements and religiously tracks each plant for anything even slightly out of the ordinary, immediately investigates whatever turns up, and informs all its other plants of whatsoever anomalies.

Such methods are not more widely employed because all besides many messengers—even the most senior executives—remain reluctant to convey bad news to bosses and colleagues. One senior executive I know in a large consumer products company had grave reservations about a takeover that was already in the works when he joined the management squad. Only, overly witting of his newcomer status, he was silent during discussions in which all the other executives seemed enthusiastic almost the plan. Many months later on, when the takeover had clearly failed, the team gathered to review what had happened. Aided by a consultant, each executive considered what he or she might have done to contribute to the failure. The newcomer, openly apologetic near his past silence, explained that others' enthusiasm had fabricated him unwilling to be "the skunk at the picnic."

In researching errors and other failures in hospitals, I discovered substantial differences across patient-care units in nurses' willingness to speak up near them. It turned out that the behavior of midlevel managers—how they responded to failures and whether they encouraged open up discussion of them, welcomed questions, and displayed humility and curiosity—was the cause. I have seen the aforementioned blueprint in a wide range of organizations.

A horrific instance in betoken, which I studied for more than two years, is the 2003 explosion of the Columbia space shuttle, which killed seven astronauts (see "Facing Ambiguous Threats," by Michael A. Roberto, Richard M.J. Bohmer, and Amy C. Edmondson, HBR Nov 2006). NASA managers spent some 2 weeks downplaying the seriousness of a piece of foam's having broken off the left side of the shuttle at launch. They rejected engineers' requests to resolve the ambivalence (which could have been done past having a satellite photograph the shuttle or asking the astronauts to conduct a space walk to inspect the area in question), and the major failure went largely undetected until its fatal consequences 16 days later. Ironically, a shared simply unsubstantiated conventionalities among program managers that there was little they could exercise contributed to their inability to find the failure. Postevent analyses suggested that they might indeed accept taken fruitful action. Simply conspicuously leaders hadn't established the necessary culture, systems, and procedures.

One challenge is teaching people in an system when to declare defeat in an experimental grade of action. The human trend to hope for the best and endeavour to avoid failure at all costs gets in the way, and organizational hierarchies exacerbate it. Equally a upshot, declining R&D projects are often kept going much longer than is scientifically rational or economically prudent. We throw expert coin after bad, praying that we'll pull a rabbit out of a hat. Intuition may tell engineers or scientists that a project has fatal flaws, only the formal decision to telephone call it a failure may be delayed for months.

Over again, the remedy—which does not necessarily involve much time and expense—is to reduce the stigma of failure. Eli Lilly has washed this since the early 1990s by holding "failure parties" to award intelligent, loftier-quality scientific experiments that fail to accomplish the desired results. The parties don't cost much, and redeploying valuable resource—specially scientists—to new projects before rather than later can salve hundreds of thousands of dollars, not to mention kickstart potential new discoveries.

Analyzing Failure

Once a failure has been detected, it's essential to go beyond the obvious and superficial reasons for it to empathise the root causes. This requires the discipline—better nonetheless, the enthusiasm—to employ sophisticated assay to ensure that the right lessons are learned and the right remedies are employed. The task of leaders is to meet that their organizations don't just move on after a failure but stop to dig in and discover the wisdom contained in it.

Why is failure analysis often shortchanged? Because examining our failures in depth is emotionally unpleasant and can chip away at our self-esteem. Left to our ain devices, most of us volition speed through or avoid failure analysis birthday. Another reason is that analyzing organizational failures requires inquiry and openness, patience, and a tolerance for causal ambiguity. Nevertheless managers typically admire and are rewarded for decisiveness, efficiency, and activeness—not thoughtful reflection. That is why the right culture is and then important.

The challenge is more than emotional; information technology's cerebral, too. Fifty-fifty without significant to, we all favor evidence that supports our existing beliefs rather than culling explanations. We also tend to downplay our responsibility and place undue arraign on external or situational factors when we neglect, only to do the reverse when assessing the failures of others—a psychological trap known equally fundamental attribution fault.

My research has shown that failure assay is frequently limited and ineffective—even in complex organizations similar hospitals, where human lives are at stake. Few hospitals systematically analyze medical errors or process flaws in guild to capture failure's lessons. Contempo research in North Carolina hospitals, published in November 2010 in the New England Journal of Medicine, found that despite a dozen years of heightened awareness that medical errors outcome in thousands of deaths each year, hospitals have non become safer.

Fortunately, there are shining exceptions to this design, which continue to provide hope that organizational learning is possible. At Intermountain Healthcare, a system of 23 hospitals that serves Utah and southeastern Idaho, physicians' deviations from medical protocols are routinely analyzed for opportunities to amend the protocols. Assuasive deviations and sharing the data on whether they really produce a better outcome encourages physicians to buy into this plan. (Meet "Fixing Health Care on the Front Lines," by Richard One thousand.J. Bohmer, HBR April 2010.)

Motivating people to go beyond first-club reasons (procedures weren't followed) to understanding the 2d- and tertiary-order reasons can be a major challenge. One way to do this is to use interdisciplinary teams with diverse skills and perspectives. Complex failures in particular are the result of multiple events that occurred in dissimilar departments or disciplines or at unlike levels of the organisation. Agreement what happened and how to prevent information technology from happening again requires detailed, team-based discussion and analysis.

A squad of leading physicists, engineers, aviation experts, naval leaders, and fifty-fifty astronauts devoted months to an analysis of the Columbia disaster. They conclusively established non only the first-social club cause—a piece of foam had hit the shuttle'south leading border during launch—but also second-order causes: A rigid bureaucracy and schedule-obsessed culture at NASA made it especially difficult for engineers to speak upwards almost anything merely the almost rock-solid concerns.

Promoting Experimentation

The third critical activity for effective learning is strategically producing failures—in the right places, at the right times—through systematic experimentation. Researchers in bones science know that although the experiments they conduct will occasionally consequence in a spectacular success, a large pct of them (70% or higher in some fields) will neglect. How practise these people go out of bed in the morning time? First, they know that failure is not optional in their work; it'south office of beingness at the leading edge of scientific discovery. Second, far more virtually of united states of america, they understand that every failure conveys valuable information, and they're eager to get it earlier the competition does.

In dissimilarity, managers in charge of piloting a new product or service—a classic case of experimentation in business—typically do whatever they can to brand sure that the pilot is perfect right out of the starting gate. Ironically, this hunger to succeed can later inhibit the success of the official launch. Too often, managers in charge of pilots design optimal conditions rather than representative ones. Thus the pilot doesn't produce noesis nigh what won't piece of work.

As well oftentimes, pilots are conducted nether optimal conditions rather than representative ones. Thus they can't show what won't work.

In the very early days of DSL, a major telecommunications company I'll call Telco did a full-scale launch of that high-speed technology to consumer households in a major urban market. It was an unmitigated client-service disaster. The visitor missed 75% of its commitments and found itself confronted with a staggering 12,000 belatedly orders. Customers were frustrated and upset, and service reps couldn't even brainstorm to answer all their calls. Employee morale suffered. How could this happen to a leading visitor with high satisfaction ratings and a make that had long stood for excellence?

A small and extremely successful suburban pilot had lulled Telco executives into a misguided conviction. The trouble was that the airplane pilot did not resemble real service conditions: It was staffed with unusually personable, skilful service reps and took identify in a customs of educated, tech-savvy customers. But DSL was a brand-new technology and, unlike traditional telephony, had to interface with customers' highly variable habitation computers and technical skills. This added complexity and unpredictability to the service-delivery challenge in means that Telco had non fully appreciated earlier the launch.

A more useful airplane pilot at Telco would take tested the technology with limited back up, unsophisticated customers, and former computers. It would take been designed to detect everything that could go wrong—instead of proving that under the best of conditions everything would get right. (Encounter the sidebar "Designing Successful Failures.") Of form, the managers in charge would have to have understood that they were going to be rewarded not for success but, rather, for producing intelligent failures as quickly equally possible.

In short, exceptional organizations are those that go beyond detecting and analyzing failures and endeavour to generate intelligent ones for the express purpose of learning and innovating. Information technology's not that managers in these organizations enjoy failure. Just they recognize it as a necessary by-product of experimentation. They also realize that they don't take to do dramatic experiments with large budgets. Oft a minor pilot, a dry out run of a new technique, or a simulation will suffice.

The backbone to confront our own and others' imperfections is crucial to solving the apparent contradiction of wanting neither to discourage the reporting of problems nor to create an environment in which anything goes. This means that managers must ask employees to be brave and speak up—and must not respond by expressing acrimony or potent disapproval of what may at first appear to be incompetence. More than often than we realize, complex systems are at piece of work behind organizational failures, and their lessons and improvement opportunities are lost when chat is stifled.

Savvy managers sympathise the risks of unbridled toughness. They know that their ability to find out almost and help resolve problems depends on their ability to larn about them. But most managers I've encountered in my research, teaching, and consulting work are far more sensitive to a different gamble—that an understanding response to failures volition simply create a lax work environment in which mistakes multiply.

This common worry should exist replaced by a new paradigm—i that recognizes the inevitability of failure in today'southward complex work organizations. Those that catch, right, and larn from failure before others do will succeed. Those that wallow in the blame game volition non.

A version of this article appeared in the Apr 2011 consequence of Harvard Business Review.