Episode Transcript
In early 2007, Carole Hemmelgarn’s life was forever changed by a failure, a tragic medical failure. At the time, she was working for Pfizer, the huge U.S. pharmaceutical firm. So she was familiar with the healthcare system. But what changed her life wasn’t a professional failure; this was personal.
Carole HEMMELGARN: My nine-year-old daughter, Allysa, was diagnosed with leukemia, A.L.L., on a Monday afternoon, and she died 10 days later. In this day and age of healthcare, children don’t die of leukemia in nine days. She died from multiple medical errors. She got a hospital-acquired infection, which we know today can be prevented. She was labeled. And when you attach labels to patients, a bias is formed, and it’s often difficult to look beyond that bias. So, one of the failures in my daughter’s care is that she was labeled with anxiety. The young resident treating her never asked myself or her father if she was an anxious child, and she wasn’t. What happens is we treat anxiety, but we don’t treat scared, afraid, and frightened. And that’s what my daughter was. Hospitals are frightening places to children.
So my daughter, with her hospital-acquired infection, became septic. But they were not treating her for the sepsis, because all they could focus on is they thought she was anxious, and they kept giving her drugs for anxiety, even though the signs, the symptoms, and me as her mother kept telling them something was wrong, something wasn’t right. They wouldn’t listen to me. So by the time, by the time she was failing so poorly and rushed to surgery and brought back out, there was nothing they could do for her. The first harm was unintentional, that they did to our daughter. It was all of the intentional harms after that, where we were lied to. The medical records were hidden from us, people were told not to talk to us. And the fact that it took the organization three years, seven months and 28 days to have the first honest conversation with us, those were all intentional harms. And that’s why in healthcare, we have to have transparency, because how many other children suffered, because of the learning that didn’t take place.
Hemmelgarn says she filed a claim against the hospital but she didn’t move forward with a lawsuit, because of the emotional toll. She ultimately took a different path: in 2021, she co-founded an advocacy group called Patients for Patient Safety U.S. It is aligned with the World Health Organization. She also runs a master’s program at Georgetown University, called Clinical Quality, Safety, and Leadership.
HEMMELGARN: When harm does reach the patient or family, that is the time to really analyze what happened. And while you never want to harm a patient or family, one of the things you’ll hear from patients and families after they have been harmed is they want to make sure that what happened to them or their loved one never happens again. The example I can give for myself personally is, I did go back to the very organization where my daughter died, and I have done work there.
Today, on Freakonomics Radio, we continue with the series we began last week, on failure. We acknowledged that some failure is inevitable:
Amy EDMONDSON: We are, by definition, fallible human beings, each and every one of us.
And that failure can be painful:
Gary KLEIN: I don’t think we should enjoy failure. I think failure needs to burn on us.
This week, we focus on the healthcare system, where failure is literally a matter of life or death.
HEMMELGARN: Some organizations felt like they had already achieved the patient-safety mission. Others, it wasn’t even part of their strategic plan.
And we will learn where, on a spectrum, to place every failure — from inexcusable …
John VAN REENEN: There’s lots of examples of huge public-sector failures. But this was one of the biggest.
To life-saving.
Bob LANGER: I really believed that if we could do this, it would make a big difference in medicine.
“How to Succeed at Failing,” Part 2, beginning now.
* * *
The story of Carole Hemmelgarn’s daughter is tragic — a hospital death caused by something other than the reason the patient was in the hospital. Unfortunately, that type of death is not as rare as you might think. Consider the case of RaDonda Vaught, a nurse at Vanderbilt University’s medical center. In 2019, she was prosecuted for having administered the wrong medication to a patient, who subsequently died. The patient was a 75-year-old woman who had been admitted to the hospital for a subdural hematoma, or bleeding in the brain. Here is RaDonda Vaught testifying at her trial:
RaDonda VAUGHT: I was pulling this medication. I didn’t think to double-check what I thought I had pulled from the machine. I used the override function. I don’t recall ever seeing any warnings that showed up on the monitor.
The medication that Vaught meant to pull from the Accudose machine was a sedative called Versed. What she mistakenly pulled was a paralytic called vecuronium. Vecuronium instead of Versed.
VAUGHT: I won’t ever be the same person. Um … It’s really … I … When I started being a nurse, I told myself that I wanted to take care of people the way that I would want my grandmother to be taken care of.
RaDonda Vaught was convicted of negligent homicide and “gross neglect of an impaired adult”; her sentence was three years’ probation. You might expect a patient-safety advocate like Carole Hemmelgarn to celebrate Vaught’s prosecution. But she doesn’t.
HEMMELGARN: This doesn’t solve problems. All this does is it creates silence and barriers. When errors happen, so often, the front line workers — your nurses, allied health physicians — were blamed. But what we’ve come to realize is, it’s really a systemic problem. They happen to be at the front line, but it’s underlying issues that are at the root of these problems. It can be policies that aren’t the right policies. It could be shortages of staff. It can be equipment failures that are known at device companies, but haven’t been shared with those using the devices. It can be medication errors because of labels that look similar, or drug names that are similar.
To get at the systemic problem in the Vanderbilt case, Hemmelgarn’s advocacy group filed a complaint with the Office of Inspector General in the Department of Health and Human Services.
HEMMELGARN: What we found most frustrating was the lack of leadership from Vanderbilt. Leadership never came out and took any responsibility. They never said anything. They never talked to the community. It was essentially silence from leadership. I think one of the other big failures we have in healthcare is fear. Healthcare is rooted in fear because of the fear of litigation. When there’s a fear of litigation, silence happens. And until we flip that model, we’re going to continue down this road.
EDMONDSON: I absolutely share that worry. And that case was, in my mind, a classic case of a complex failure. Yes, there was a human error. We also had faulty medication-labeling and storing practices, with alphabetical organization of drugs, which is not how you do it.
That’s Amy Edmondson. We heard from her last week too. She’s an organizational psychologist at the Harvard Business School. She recently published a book called Right Kind of Wrong: The Science of Failing Well. The Vanderbilt case was not an example of failing well. Radonda Vaught, you’ll remember, dispensed Vecuronium instead of Versed.
EDMONDSON: You know, you don’t have a dangerous, potentially fatal drug next to one that’s routinely used in a particular procedure. It’s what we might call an accident waiting to happen. With that perspective in mind, RaDonda is as much a victim of a system failure as a perpetuator of the failure, right? So, this reaction, to make this a criminal — human error is almost never criminal. To criminalize this I think reflects an erroneous belief that by doing so will preclude human error. No. What we will do is preclude speaking up about human error. And to her credit, she spoke up — and that, one could argue, ultimately led her to the conviction. She would have been better off somehow trying to hide it, which I wouldn’t advocate, obviously. But when we recognize deeply recognize that errors will happen, then that means that what excellence looks like is catching and correcting errors and then being forever on the lookout for vulnerabilities in our systems.
Let’s take a step back and look at the scale of this problem. In 1999, the Institute of Medicine — known today as the National Academy of Medicine — issued a report called “To Err Is Human: Building a Safer Health System.” It found that 2 to 3 percent of all U.S. hospital admissions result in preventable injury or death, with medical error causing between 44,000 and 98,000 deaths per year. A more recent study, published in 2013 in the Journal of Patient Safety, put the number of preventable deaths at U.S. hospitals at more like 200,000 a year. These large numbers of preventable deaths got a lot of attention in the medical community — but Carole Hemmelgarn says the attention hasn’t produced enough change.
HEMMELGARN: Some organizations felt like they had already achieved the patient-safety mission. Others, it wasn’t even part of their strategic plan. There’s areas where improvement has definitely escalated since the report came out over 20 years ago. But it hasn’t been fast enough. What we see is that not everything is implemented in the system, that you can oftentimes have champions that are doing this work — and if they leave, the work isn’t embedded and sustainable.
Amy Edmondson at Harvard has been doing research on medical failure for a long time. But she didn’t set out to be a failure researcher.
EDMONDSON: As an undergraduate, I studied engineering sciences and design.
DUBNER: Tell me about the first phase of your professional life, including with Buckminster Fuller.
EDMONDSON: So I’m answering that question with a huge smile on my face. I worked three years for Buckminster Fuller, who was an octogenarian creative person, an inventor, a genius, a writer, a teacher — best-known for the geodesic dome, which he invented, but single-mindedly about: how do we use design to make a better world? You can’t get people to change; you have to change the environment, and then they’ll change with it, was a kind of notion that he had. My part was just doing engineering drawings and building models and doing the mathematics behind new, simpler, geodesic configurations. And it was so much fun.
DUBNER: And what was his view on failure generally?
EDMONDSON: Ah! He was a very enthusiastic proponent of using failure to learn. He said often the only mistake we make is thinking we shouldn’t make mistakes. He would give the example of the very first time he got a group of students together to build a geodesic dome that he had — you know, he’d done the math, he’d come up with this idea and he got 20 students together, they’re outside, they built the thing and it immediately collapsed. And he enthusiastically said, “Okay, that didn’t work. Now, what went wrong?” And it was really the materials they were using, which were, I think, the best way to describe them as Venetian blind materials. They had the tensile strength, but they certainly didn’t have the compressive strength to do their job.
DUBNER: Okay, and what are the steps you take to turn that failure into a useful “learning,” I guess is the noun we use these days.
EDMONDSON: Immediate diagnosis, right? We gotta step back, “Okay, what did we set out to do? What actually happened? Why might that be the case? What do we do differently next time? I mean, that’s a sort of a rough outline of an after-action review. It could be flawed assumptions, it could be flawed calculations. It could be any number of things. And we don’t know until we put our heads together and try to figure it out.
It was several years into her engineering career that Edmondson decided to get a Ph.D. in organizational behavior.
EDMONDSON: I was interested in learning in organizations, and I got invited to be a member of a large team studying medication errors in hospitals. And the reason I said “yes” was, first of all, I was a first-year graduate student. I needed to do something. And second of all, I saw a very obvious link between mistakes and learning. And so I thought, here we’ve got these really smart people who will be identifying mistakes. And then I can look at how do people learn from them, and how easy is it and how hard is it. So that’s how I got in there. And then one thing led to another — after doing that study, people kept inviting me back.
DUBNER: I see. “She loves failure, they say.”
EDMONDSON: That’s right.
Edmondson focused her research on what are called preventable adverse drug events — like the one from the RaDonda Vaught case.
EDMONDSON: Now, you can divide adverse drug events into two categories: one which is related to some kind of human error or system breakdown, and the other, which is a previously unknown allergy, so literally couldn’t have been predicted. And those are still adverse drug events, but they’re not called preventable adverse drug events.
DUBNER: But within the first category, there’s probably ten subcategories at least, right? There’s bad data-entry, bad handwriting, wrong eyeglasses.
EDMONDSON: On and on it goes, yup. Or, you know, using language badly so that people didn’t understand what you said, and they didn’t feel safe asking.
DUBNER: My wife had a knee surgery, easy knee surgery, and the painkiller that they prescribed on the spot — the doc actually stood there and wrote it — was for 100x the dosage.
EDMONDSON: See, that’s an error-driven, preventable adverse drug event.
DUBNER: Yes, I agree.
EDMONDSON: You know, there will always be things that go wrong, or at least not the way we wanted them to. In my observation, in studying teams in a variety of industries and settings, was that responses to failure were rather uniform. Inappropriately uniform! The natural response, and even the formal response, was to find the culprit. As if there was a culprit. And either discipline or retrain or, you know, shame and blame the culprit. And it wasn’t a very effective solution because the only way to prevent those kinds of system breakdowns is to be highly vigilant to how little things can line up and produce failures.
Based on what she was learning from medical mistakes, Edmondson wanted to come up with a more general theory of failure — or, if not a theory, at least a way to think about it more systematically. To remove some of the blame. To make the responses to failure less uniform. Over time, she produced what she calls — well, here, let’s have Edmondson say it:
EDMONDSON: My spectrum of causes of failures.
Coming up, we’ll hear about that spectrum of causes of failures. It can clarify some things — but not everything.
EDMONDSON: Uncertainty is everywhere.
* * *
By the way, if you consider yourself a super-fan of this show, we’ve just launched a new membership program, Freakonomics Radio Plus. Every week, members get a bonus episode of Freakonomics Radio; you also get to listen ad-free to this show and the other shows in the Freakonomics Radio Network. To sign up, visit the Freakonomics Radio show page on Apple Podcasts, or go to freakonomics.com/plus. If you don’t want to become a member, just do nothing. Everything will stay the same.
* * *
How did Amy Edmondson become so driven to study failure? Well, here’s one path to it. Her whole life, she had been a straight-A student.
EDMONDSON: I never had an A-minus. Well, you know, I once had one in 10th grade. It just was so devastating, I resolved not to have one again. And I’m only partly joking.
But then she went to college.
EDMONDSON: I got an F on my first-semester multivariable calculus exam. An F. Like, I failed the exam. I mean, that’s unheard of.
DUBNER: What’d that feel like?
EDMONDSON: I didn’t see it coming. But I wasn’t baffled after the fact. After the fact, it was very clear to me that I hadn’t studied enough.
In the years since then, Edmondson has been refining what she calls “a spectrum of causes of failure.” The spectrum ranges from “blameworthy” to “praiseworthy,” and it contains six distinct categories of failure.
EDMONDSON: Let’s take two extremes. Let’s say something goes wrong, we achieve an undesired result. On one end of the spectrum, it’s sabotage, someone literally tanked the process. They threw a wrench into the works. On the other end of the spectrum, we have a, a scientist or an engineer hypothesizing some new tweak that might solve a really important problem and they try it, and it fails. And of course we praise the scientist and we punish the saboteur. But the gradations in between often lull us into a false sense that it’s blameworthy all the way.
Okay, so let’s start at the blameworthy end of the spectrum, and move our way along. Number one, of the six:
EDMONDSON: My spectrum of causes of failures starts with sabotage or deviance. I soak a rag in lighter fluid, set it on fire, or throw it into a building, right? Or I’m a physician in a hospital, I am a surgeon, and I come to work drunk, and do an operation.
DUBNER: You describe this as “the individual chooses to violate a prescribed process or practice.” Now, I could imagine there are some cases where people violate because they think that the process is wrong.
EDMONDSON: That’s right. There has to be intent here. To label something a true sabotage, it has to be: my intent is to break something. It’s not a mistake and it’s not a thoughtful experiment. There certainly are protocols in hospitals, for example, where thoughtful physicians will deliberately depart from the protocol because their clinical judgment suggests that would be better. They may be right, they may be wrong, but that would not qualify as a blameworthy act.
After sabotage on the spectrum comes inattention.
EDMONDSON: Inattention is when something goes wrong because you just were mailing it in, you spaced out, you didn’t hear what someone said, and you didn’t ask, and then you just tried to wing it. Or you maybe are driving, you’re a trucker and you’re driving, and you look away or fiddle with the radio and have a car crash.
DUBNER: Now, it sounds like those are mostly blameworthy, but what about inattention caused by external factors?
EDMONDSON: Well, that’s exactly right. Once we leave sabotage and move to the right in the spectrum, it will never be immediately obvious whether something’s blameworthy or not. It’s always going to need further analysis. So when we say the failure was caused by someone not paying attention, that just brings up more questions. Okay, why weren’t they paying attention? Now, it could be that this poor nurse was on a double shift, and that is not necessarily the nurse’s fault. It might be the nurse manager who assigned that double shift. Or it might be the fact that someone else didn’t show up. And so they have to just do it, and they’re quite literally too tired to pay attention fully, right? So we always want to say, “Well, wait, let’s see, what are the other contributing factors to this inattention?”
DUBNER: Can you think of a large-scale failure, a corporate or institutional failure, that was caused largely by inattention?
EDMONDSON: Yes. One that comes to mind is a devastating collapse with the loss of many lives, when a Hyatt Regency atrium collapsed in Kansas City in the early eighties. And the inattention there was the engineer-on-record’s failure to pay close attention when the builder decided — out loud, not hidden — to swap one long beam for two smaller connected steel beams. It would have been a five-minute calculation to show that won’t work with the loads that were expected. It was a change that didn’t obtain the attention it needed to have avoided this catastrophic failure.
DUBNER: And was that change done to save money, or was it even more benign than that?
EDMONDSON: I think it was a combination of speed and money. Speed is money.
DUBNER: Wow, wow, wow, wow. That’s a great example. Okay, let’s go to the next one: inability. I’m reading one version of your spectrum here, which describes this as “the individual lacks the knowledge, attitudes, skills, or perceptions required to execute a task.” That’s quite a portfolio of potential failure.
EDMONDSON: That’s right. And that spans from a young child who doesn’t yet know how to ride a bicycle, so as soon as they hop on that bicycle, they’re going to fall off, because they don’t have the ability yet, to multivariable calculus, which at least when you’re not studying hard enough, you don’t have the ability. So it’s something that you just don’t have the ability to do to success, but usually could develop.
DUBNER: This reminds me of the Peter Principle, where people get promoted to a position higher than they’re capable, based on their past experience. But their past experience may not have been so relevant to this.
EDMONDSON: That’s a great connection, yeah, the Peter Principle, where the failure gets caused by the fact that you don’t have the ability to do the new role, but no one really paused to reflect on that.
DUBNER: I sometimes think about this in the political realm, too. The ability to get elected and the ability to govern effectively seem to be almost uncorrelated to me, I’m sorry to say. Do you think that’s the case? And do you apply this spectrum sometimes to the political realm?
EDMONDSON: I don’t think it was always the case, but I think it might be increasingly the case. There’s no theoretical reason why the two abilities — to be compelling and win people over to your point of view — should be at odds with the capability to do it. But the way it is increasingly set up in our society might be putting them at odds.
After inability comes what Edmondson calls “task challenge.”
EDMONDSON: Yes, the task is too challenging for reliable, failure-free performance.
DUBNER: Example?
EDMONDSON: A great example is an Olympic gymnast, who is training all the time and is able to do some of the most challenging maneuvers, but will not do them 100 percent of the time. And so when that person experiences a failure — they trip during their routine — then we would call that a failure that was largely caused by the inherent challenge of the task.
DUBNER: Can you give an example in either the corporate or maybe academic realm?
EDMONDSON: Let’s go to NASA, for example. The shuttle program is very, very challenging. I think we can all agree to that. And over time, they started to think of it as not challenging. But really, it’s a remarkably challenging thing to send a rocket into space and bring it back safely.
DUBNER: Kind of paradoxical then that the thing was actually called Challenger.
EDMONDSON: Oh, that’s a good point. Actually, I love Richard Feynman. Looking back on the Challenger accident, his sort of simple willingness to just put the piece of O-ring in the ice water, see what happens, right? That’s something that, in a better-run, more psychologically safe, more creative, generative work environment, someone else would have done real-time.
DUBNER: But if I recall correctly, even though he was on that commission to investigate, they tried to essentially shut him up. They didn’t want that news coming out at the hearing. They didn’t want the failure to be so explicit.
EDMONDSON: That’s right. But that’s — I mean, that’s not a good thing.
DUBNER: That’s not a good thing.
EDMONDSON: You’ve got to learn from it, so that it doesn’t happen again.
By the way, if you don’t remember the story of Richard Feynman and the Challenger investigation and the O-rings — don’t worry. We’re working on a show about Feynman that you’ll hear in the coming months. Okay, back to failure: the fifth cause of failure on Amy Edmondson’s spectrum is uncertainty.
EDMONDSON: So, uncertainty is everywhere. There’s probably an infinite number of examples here. But let me pick a silly one. A friend sets you up on a blind date. And you like the friend, and you think, okay, sure. And then you go out on the date and it’s a terrible bore or worse, right? It’s a failure. But you couldn’t have known in advance, it was uncertain.
DUBNER: How about a less-silly example?
EDMONDSON: You’re in a company setting. You have an idea for a strategic shift or a product that you could launch. And there’s very good reasons to believe this could work. But it’s not 100 percent.
The final cause of failure — we have by now moved all the way from the blameworthy end of the spectrum to the praiseworthy — is simply called experimentation.
EDMONDSON: I’m being fairly formal when I say experimentation, right? The most obvious example is a scientist in a lab and probably really believes it will work and puts the chemicals in and lo and behold, it fails. Or in much smaller scale I’m going to experiment with being more assertive in my next meeting, and doesn’t quite work out the way I’d hoped. It’s the Edison quote, you know, 10,000 ways that didn’t work. He’s perfectly, perfectly willing to share that because he’s proud of each and every one of those 10,000 experiments.
So that is Amy Edmondson’s entire spectrum of the causes of failure: sabotage, inattention, inability, task challenge, uncertainty, and experimentation. If you’re like me, as you hear each of the categories you automatically try to match them up with specific failures of your own. If nothing else, you may find that thinking about failure on a spectrum, from blameworthy to praiseworthy, is more useful than the standard blaming-and-shaming. It may even make you less afraid of failure. That said, not everyone is a fan of Edmonson’s ethos of embracing failure. A research article by Jeffrey Ray at the University of Maryland, Baltimore County is called “Dispelling the Myth that Organizations Learn from Failure.” He writes, “Failure shouldn’t even be in a firm’s vocabulary. To learn, from failure or otherwise, a firm must have an organizational learning capability. If the firm has the learning capability in the first instance, why not apply it at the beginning of a project to prevent a failure, rather than waiting for a failure to occur and then reacting to it.” But Amy Edmonson’s failure spectrum has been winning admirers — including Gary Klein, the research psychologist best known as the pioneer of naturalistic decision-making.
Gary KLEIN: I’m very impressed by it. I’m impressed because it’s sophisticated. It’s not simplistic. There’s a variety of levels, and a variety of reasons. And before we start making policies about what to do about failure, we need to look at things like her spectrum and identify what kind of a failure is it so that we can formulate a more effective strategy.
Okay, let’s do that. Coming up: two case studies of failure, one of them toward the blameworthy end of the spectrum.
John VAN REENEN: It was very much driven by the prime minister, Tony Blair.
The other, quite praiseworthy.
Bob LANGER: I failed over 200 times before I finally got something to work.
* * *
John Van Reenen is a professor at the London School of Economics. He studies innovation, but years ago he did some time in the British civil service.
John VAN REENEN: I spent a year of my life working in the Departments of Health when there was a big expansion in the U.K. National Health Service of resources, of various attempts at reforms.
The National Health Service is the U.K.’s publicly funded healthcare system, and there was one particular reform that Van Reenen got to see up close.
VAN REENEN: One of the key things that was thought could really be a game-changer was to have electronic patient records. So you can see the history of patients, know the conditions, what they’ve been treated with. And having that information — I mean, instead of having all these pieces of paper written illegibly by different physicians, you could actually have this in a single record — would not only make it much easier to find what was going on with patients, but could also be used as a data source to try and help think about how patients could have more joined-up care and could even maybe predict what kind of conditions they might have in the future.
The project was called Connecting for Health, and there was substantial enthusiasm for it; at least the ad campaign was enthusiastic:
ANNOUNCER: All this is a key element in the future of the N.H.S. One day, not too far away, you’ll wonder how you ever lived without it.
VAN REENEN: It was very much driven by the prime minister, Tony Blair. This was a centralized, top-down approach in order to have a single I.T. system where you could access information, instead of having all these different I.T. systems, these different siloed pieces of paper, to have it in one consistent national system.
The N.H.S. is a big operation, one of the biggest employers in the world.
VAN REENEN: I think the ranking goes something like the U.S. Department of Defense, the Indian railway system, Walmart — and the N.H.S. is up there in the top five. But then if you drill down into it, it is pretty fragmented. Each local general practitioner unit is self-employed. Each trust has a lot of autonomy. And that’s part of the issue is that this was a centralized top-down program in a system where there’s a lot of different fiefdoms. A lot of different pockets of power who are quite capable of resisting this, and disliked very strongly being told “this is what you’re going to have, this is what you’re going to do,” without really being engaged and consulted properly.
But the train rolled on, despite these potential problems. Connecting for Health required a massive overhaul of hardware systems as well as software systems.
VAN REENEN: And the delivery of those was — there was a guy called Richard Granger who was brought in, I think he was the highest-paid public servant in the country. He was at Deloitte before he came. And then after he left, he went to work for Accenture. He was brought in to do this, and he designed these contracts — very tough contracts, which loaded the risk of things going wrong very strongly onto the private-sector providers. I think just about every single “winner” eventually either went bankrupt or walked away from the contract. The estimates vary of the cost of this, but estimates are up to $20 billion lost on this project. It was it was the biggest civilian I.T. project in the Western world. I mean, there’s lots of examples of huge public-sector failures, and private-sector failures as well. But this was one of the biggest.
British Parliament ultimately called this attempted reform “one of the worst and most expensive contracting fiascos ever.” So what kind of lessons can be learned from this failure?
VAN REENEN: I think it’s a failure of many, many different causes, of many different levels. That top-downness, the not really understanding what was going on at a grassroots level, and the haste — it was attempted very quickly.
DUBNER: I’ve read that the haste — especially the haste of awarding contracts — at the time was considered a great thing, because it was so atypical of how government worked and it was hailed as a new way of the government doing business. In the end, that haste turned out to be a problem though, correct?
VAN REENEN: Correct. I mean, it seemed at the time when these contracts were formed, the government was getting a good deal. They were doing it quickly, they were loading the risks onto the suppliers. So it wasn’t obvious from the get-go that this was going to be as bad as it turned out to be. Looking back, trying to do things quickly in such a complicated system — there was so much complexity that a lot of these contracts effectively had to be rewritten afterwards. And I think another general lesson is that when you’re doing a long-term important, big contract, you can’t get everything written down quickly. There has to be a lot of give and take. It’s a kind of relationship that you have to adjust as things go. You know, contracts are very fuzzy. They’re very incomplete. You just have to accept that, that you’re going to have to not get things right, but not try to do everything really, really, really quickly. An I.T. project is never just about I.T. It’s also about the way you change a whole organization.
And to do it is not just about spending money. You also have to get players in the system on board because it’s very difficult to just get them to do things — especially in a public system where you can’t just fire people if you want to fire them. You really have to have a culture of kind of bringing people on board if you want to make these type of changes, and that just didn’t happen. So I don’t think it’s just one thing you could think of. There’s the haste, there’s the design, which worked out badly, and there’s the cultural aspects that we’ve talked about. When you’re trying to innovate, you want to have a way of allowing people to take risks and do things wrong. But then you also have to have feedback mechanisms to figure out, what has gone wrong? So, creating an attitude of saying, well, we actually don’t know what the right thing to do is so we’re prepared to do experimentations and learn from that.
If you’re the kind of person who likes to understand and analyze failure in order to mitigate future failures, what might be useful here is to overlay the National Health Service’s I.T. fiasco onto Amy Edmondson’s spectrum of causes of failure. Reconfiguring a huge I.T. system certainly qualifies as a “task challenge”; but there were shades of inability and inattention at work here as well. All of those causes reside toward the “blameworthy” end of the scale. As for the “praiseworthy” end of the spectrum — that’s where experimentation can be found. The N.H.S. project didn’t incorporate much experimentation; it was more command-and-control, top-down, with little room for adjustment and little opportunity to learn from the small failures that experimentation can produce, and which can prevent big failures. Experimentation, if you think about it, is the foundation of just about all the learning we do as humans. And yet we seem to constantly forget this. Maybe that’s because experimentation will inevitably produce a lot of failure — I mean, that’s the point! — and most of us just don’t want to fail at all, even if it’s in the service of long-term success. So let’s see if we can’t adjust our focus here. Let’s talk about real experimentation. And for that, we’ll need not another social scientist — like John Van Reenen or Amy Edmondson, as capable as they are — but an actual science scientist. Here is one of the most acclaimed scientists of the modern era.
LANGER: My name’s Bob Langer, and I’m an institute professor at M.I.T. I do research, but I’ve also been involved in helping get companies started. And I’ve done various advising to the government — F.D.A., and places like that.
DUBNER: And if I say to you, “Bob, what kind of scientist are you exactly?” how do you answer that question?
LANGER: Well, I would say I’m a chemical engineer or a biomedical engineer, but people have called me all kinds of things. You know, they’ve called me a biochemist. We do very interdisciplinary work, so I end up getting called more than one thing.
DUBNER: Do you care what people call you?
LANGER: I just like them to call me Bob.
Langer holds more than 1,400 patents, including those that are pending. He runs the world’s largest biomedical engineering lab, at M.I.T., and he’s one of the world’s most highly-cited biotech researchers. He also played a role in the founding of dozens of biotech firms, including Moderna, which produced one of the most effective Covid vaccines. One thing Langer is particularly known for is drug delivery: that is, developing and refining how a given drug is delivered and absorbed at the cellular level. A time-release drug, for instance, is the sort of thing we take for granted today. But it took a while to get there. One problem Langer worked on back in the 1970s was finding a drug-delivery system that would prevent the abnormal growth of blood vessels. The chemical that inhibits the growth is quite large by biological standards, and there was consensus at the time that a time-release wouldn’t work on large molecules. But as Langer once put it, “I didn’t know you couldn’t do it because I hadn’t read the literature.” So, he ran experiment after experiment after experiment before finally developing a recipe that worked. Decades later, thanks to all that failure, his discovery played a key role in how Moderna used messenger RNA to create its Covid vaccine.
DUBNER: So, in your line of work, when I say the word “failure,” what comes to mind?
LANGER: Well, a lot of things, but I’d go back to my own career. I failed at trying to get research grants. My first nine research grants were turned down. I sent them to places like the National Institutes of Health, and they have study sections, reviewers. Mine would go, just because of the work I was doing, to what was called a Pathology B study section. And they would review it. And they said, “Well, Dr. Langer, he’s an engineer. He doesn’t know anything about biology or cancer.” I failed over and over again. Other things, like, I failed to get a job in a chemical engineering department, as an assistant professor even. Nobody would hire me. They said, actually, the opposite. They said, “Chemical engineers don’t do experimental biomedical engineering work. You know, they should work on oil or energy.” When I first started working on creating these micro- or nanoparticles to try to get large molecules to be delivered, I failed over 200 times, I mean, before I finally got something to work. I could go on and on on my failures.
DUBNER: What kept you going during all this failure?
LANGER: I really believed that if we could do this, it would make a big difference in science, and I hoped a big difference in medicine. Secondly, as I did some of it, I could see some of these results with my own eyes. You know, when we were trying to deliver some of these molecules to stop blood vessel growth, I could see. We were doing this double-blind, but I could still see that we were stopping the vessels from growing. That’s such a visual thing. And I also developed these ways of studying delivery out of the little particles by putting certain enzymes in them and putting dyes in a little gel that would turn color if the enzymes came out. And I could see that happening. Like I say, the first 200 times or first 200 designs or more, it didn’t happen. But finally I came up with a way where I’d see it come out after an hour, after two hours, after a day, after a second day, up to over 100 days, in some cases. So I could see with my own eyes this was working. So that made an enormous difference to me too.
DUBNER: But failing 200 times costs a lot of money and obviously a lot of time. Did you ever almost run out of one or the other?
LANGER: The experiments I was doing weren’t that expensive, especially the delivery ones initially, because they were in test tubes. I worked probably 20-hour days. And so the expense wasn’t that great. And I’ve always been good at manufacturing time.
DUBNER: Now, let’s say someone is in a similar situation today to where you were then, with an idea or a set of ideas that they believe in, that they think they are right about, they think it’s an important idea, and yet they are failing and failing to get the attention of the people who can help manufacture success. How do you think about the line — I think of it sometimes as the line between grit and quit, right? Economists talk about opportunity cost. Every hour you spend on something that isn’t working is an hour you could spend on something that is working. But then psychologists talk about grittiness, and how useful it can be to stick things out. Do you have anything to say to people who might be wrestling with that?
LANGER: Well, I think it’s a great question. And I ultimately think it’s a judgment call. And we can never be sure of our judgment. You like to try to think, are these things scientifically possible? I think that’s one thing. Secondly, it’s good to get advice from people. It doesn’t mean you have to take it, but it’s good to get advice. I certainly personally have always erred on the side of, I guess, not quitting. And maybe that’s sometimes a mistake. I don’t think so, I think it depends on what could happen if you are successful. You know, if you are successful, could it make a giant difference in the world? Could it help science a lot? Could it help patients’ lives a lot? And so, if you really feel that it can, you try that much harder. If it’s incremental, sure. Then it’s much easier to quit.
DUBNER: Is that ability to persevere — within yourself at least — do you think that’s your natural temperament? Is that something you learned? Did you find incentives to lead you there?
LANGER: I think for me, there were a couple of things. One, I guess I’ve always been very stubborn. My parents told me that. But secondly, I think there’s a whole thing with role models, too. When I was a postdoc, the man that I worked with, Judah Folkman, he experienced the same thing. He had this theory that if you could stop blood vessels, you could stop cancer, and that was mediated by chemical signals. And everyone told him he was wrong. But I would watch him every day, and he believed anything was possible. And he kept sticking to it. And, of course, eventually he was right. I think seeing his example probably also had a big effect on me.
DUBNER: Can you talk to me about how scientific failure is treated generally? Let’s assume a spectrum. And on one end of the spectrum is that every failure is written up and published, and perhaps even celebrated as having discovered a definitive wrong path to pursue. So everybody coming after you can cross that off their list. And on the other end of the spectrum, every failure is hidden away, which allows many other people to make the same failure. Can you talk about where the reality is?
LANGER: I think that’s an interesting question. A lot of it even depends how you define failure. You know, when you’re trying to learn about something, you try different things. And embedded in the scientific papers we write — like when we wrote this paper in Nature, in 1976, which was the first time you could get small particles to release large molecules from biocompatible materials — well, some of the materials we used failed. A lot of them did, actually, because they would either cause inflammation or the drug would come out way too fast or not come out at all. We found one fraction that worked, and stopped blood vessels, and probably 50 or 100 that didn’t. So the failures and successes are maybe in the same papers sometimes. What I’ve tried to do, even to give more detail, is you put all the data in, even if it makes for a very long thesis. So not only are the graphs there in the papers, but there’s even the raw data that people can look at and analyze. And I try to get people to do as much of that as possible. So I guess what I’m trying to say is that the failures and successes are almost intertwined.
DUBNER: I’d like to hear you talk about how failure is discussed or thought of in the lab. Maybe it’s nothing overt. But I am curious, especially when you bring in young people, researchers, whether they’re, you know, postdoc or undergrad, do you give pep talks about failure? Do you kind of have a philosophy that you want to instill in these people that failure is an essential component of research and success?
LANGER: Yes. And I do. Whether it’s my own talks or just meeting with students and, brainstorming with them about those things. But to me, that research, scientific research, I mean, you just fail way more than — at least I do, way more than you succeed. It’s just part of the process. I mean, that’s experimentation. And that’s okay.
DUBNER: A lot of your colleagues and students go on to start companies, and that’s a whole different ball of wax. How do you think about failure in the entrepreneurial process?
LANGER: Obviously the easy criteria is a successful company, having a good financial exit, I suppose. But I don’t necessarily think of it as just that way. I mean, that’s certainly going to be important. You know, I’ve been involved in things where you’ve advanced science and you learn some things, and there’s degrees of success. You just don’t know. I’ve been pretty fortunate in the companies we’ve started in terms of the exits that they’ve had. But I just think there’s no simple criteria. I feel like we’ve turned out a lot of great scientists and entrepreneurs, and not all their companies have had great financial exits. But I think they’ve also created products that can change people’s lives. And that, to me, is also very, very important, obviously. That’s why we do it in the first place. I have never done it for money. And I don’t think they do it for money. They do it to try to make a difference in the world.
DUBNER: Do you think failure is, however, a different animal in the research sphere as in the entrepreneurial sphere?
LANGER: I would say yes, I think it is. But I also think there’s different cultures, too. I think the good thing about the United States culture, maybe in contrast to some cultures, is failure’s widely accepted. I’ll give you one of my examples, actually, in the business sphere. So, I’m a big fan of chocolate.
DUBNER: Of eating it or making it or researching it?
LANGER: Probably any part, but mostly eating it. But at any rate, one of the books I read — and I’m actually not a fan of their chocolate — is a book on Milton Hershey. And so, this really gets to your point on failure. Milton Hershey, he had this idea when he was young, very young, of starting a candy company. And I remember the first candy company, he went bankrupt, and he tried to raise more money, started another one. I think like the first six or seven totally failed, but not the last one, obviously. And he became a millionaire at a time when there weren’t very many. Was that really failure? Or was it just being an apprentice to trying to learn how to succeed? And I think that’s true in a lot of things. The reason I brought it up is I don’t think there’s a shame in failure in either area. Or I hope there’s not. I think you have to feel it’s okay, and then you keep going on.
What do you think? Would you like to live in a world where there’s no shame in failure? Or: do you think it’s important for failure to hurt, to burn (as one of our guests put it last week); maybe that creates a stronger incentive to succeed? I’d love to know your thoughts on this question, and on this series so far; send an email to radio@freakonomics.com, or leave a review or rating in your podcast app. Coming up next time on the show: we’ll dig deeper into the idea of grit versus quit. When you’re failing, how do you know if it’s time to move on?
John BOYKIN: We just could not stop it from leaking, and I was no longer willing to just keep pouring more and more of my money into it.
Helen FISHER: He dumped me when I was 70. And I married him again at age 75.
Travis THUL: You know, hope springs eternal. This is a great idea. You just have to raise a quarter million dollars.
Case studies in failure, and in grit versus quit — including stories from you, our listeners. That’s next time on the show. Until then, take care of yourself and, if you can, someone else too. And remember to check out Freakonomics Radio Plus if you want even more Freakonomics Radio; every week you’ll get a bonus episode of the show — this week, for instance, you’ll hear our full interview with the remarkable Bob Langer. To sign up, visit the Freakonomics Radio show page on Apple Podcasts, or go to freakonomics.com/plus.
* * *
Freakonomics Radio is produced by Stitcher and Renbud Radio. This episode was produced by Zack Lapinski and mixed by Eleanor Osborne, with help from Jeremy Johnston. Our staff also includes Alina Kulman, Elsa Hernandez, Gabriel Roth, Greg Rippin, Jasmin Klinger, Julie Kanfer, Lyric Bowditch, Morgan Levey, Neal Carruth, Rebecca Lee Douglas, Ryan Kelley, and Sarah Lilley. Our theme song is “Mr. Fortune,” by the Hitchhikers; all the other music was composed by Luis Guerra.
Sources
- Amy Edmondson, professor of leadership management at Harvard Business School.
- Carole Hemmelgarn, co-founder of Patients for Patient Safety U.S. and director of the Clinical Quality, Safety & Leadership Master’s program at Georgetown University.
- Gary Klein, cognitive psychologist and pioneer in the field of naturalistic decision making.
- Robert Langer, institute professor and head of the Langer Lab at the Massachusetts Institute of Technology.
- John Van Reenen, professor at the London School of Economics.
Resources
- Right Kind of Wrong: The Science of Failing Well, by Amy Edmondson (2023).
- “Reconsidering the Application of Systems Thinking in Healthcare: The RaDonda Vaught Case,” by Connor Lusk, Elise DeForest, Gabriel Segarra, David M. Neyens, James H. Abernathy III, and Ken Catchpole (British Journal of Anaesthesia, 2022).
- “Dispelling the Myth That Organizations Learn From Failure,” by Jeffrey Ray (SSRN, 2016).
- “A New, Evidence-Based Estimate of Patient Harms Associated With Hospital Care,” by John T. James (Journal of Patient Safety, 2013).
- To Err is Human: Building a Safer Health System, by the National Academy of Sciences (1999).
- “Polymers for the Sustained Release of Proteins and Other Macromolecules,” by Robert Langer and Judah Folkman (Nature, 1976).
- The Innovation and Diffusion Podcast, by John Van Reenen and Ruveyda Gozen.
Extras
- “How to Succeed at Failing,” series by Freakonomics Radio (2023).
- “Will a Covid-19 Vaccine Change the Future of Medical Research?” by Freakonomics Radio (2020).
- “Bad Medicine, Part 3: Death by Diagnosis,” by Freakonomics Radio (2016).
Comments