How a Deadly Fire Gave Birth to Modern Medicine

How a Deadly Fire Gave Birth to Modern Medicine

The Cocoanut Grove fire in Boston in 1942 tragically claimed 490 lives, but was the catalyst for several important medical advances.

Boston Public Library

On the evening of November 28, 1942, more than 1,000 revelers from the Boston College-Holy Cross football game jammed into the Cocoanut Grove, Boston's oldest nightclub. When a spark from faulty wiring accidently ignited an artificial palm tree, the packed nightspot, which was only designed to accommodate about 500 people, was quickly engulfed in flames. In the ensuing panic, hundreds of people were trapped inside, with most exit doors locked. Bodies piled up by the only open entrance, jamming the exits, and 490 people ultimately died in the worst fire in the country in forty years.

"People couldn't get out," says Dr. Kenneth Marshall, a retired plastic surgeon in Boston and president of the Cocoanut Grove Memorial Committee. "It was a tragedy of mammoth proportions."

Within a half an hour of the start of the blaze, the Red Cross mobilized more than five hundred volunteers in what one newspaper called a "Rehearsal for Possible Blitz." The mayor of Boston imposed martial law. More than 300 victims—many of whom subsequently died--were taken to Boston City Hospital in one hour, averaging one victim every eleven seconds, while Massachusetts General Hospital admitted 114 victims in two hours. In the hospitals, 220 victims clung precariously to life, in agonizing pain from massive burns, their bodies ravaged by infection.


The scene of the fire.

Boston Public Library

Tragic Losses Prompted Revolutionary Leaps

But there is a silver lining: this horrific disaster prompted dramatic changes in safety regulations to prevent another catastrophe of this magnitude and led to the development of medical techniques that eventually saved millions of lives. It transformed burn care treatment and the use of plasma on burn victims, but most importantly, it introduced to the public a new wonder drug that revolutionized medicine, midwifed the birth of the modern pharmaceutical industry, and nearly doubled life expectancy, from 48 years at the turn of the 20th century to 78 years in the post-World War II years.

The devastating grief of the survivors also led to the first published study of post-traumatic stress disorder by pioneering psychiatrist Alexandra Adler, daughter of famed Viennese psychoanalyst Alfred Adler, who was a student of Freud. Dr. Adler studied the anxiety and depression that followed this catastrophe, according to the New York Times, and "later applied her findings to the treatment World War II veterans."

Dr. Ken Marshall is intimately familiar with the lingering psychological trauma of enduring such a disaster. His mother, an Irish immigrant and a nurse in the surgical wards at Boston City Hospital, was on duty that cold Thanksgiving weekend night, and didn't come home for four days. "For years afterward, she'd wake up screaming in the middle of the night," recalls Dr. Marshall, who was four years old at the time. "Seeing all those bodies lined up in neat rows across the City Hospital's parking lot, still in their evening clothes. It was always on her mind and memories of the horrors plagued her for the rest of her life."

The sheer magnitude of casualties prompted overwhelmed physicians to try experimental new procedures that were later successfully used to treat thousands of battlefield casualties. Instead of cutting off blisters and using dyes and tannic acid to treat burned tissues, which can harden the skin, they applied gauze coated with petroleum jelly. Doctors also refined the formula for using plasma--the fluid portion of blood and a medical technology that was just four years old--to replenish bodily liquids that evaporated because of the loss of the protective covering of skin.

"Every war has given us a new medical advance. And penicillin was the great scientific advance of World War II."

"The initial insult with burns is a loss of fluids and patients can die of shock," says Dr. Ken Marshall. "The scientific progress that was made by the two institutions revolutionized fluid management and topical management of burn care forever."

Still, they could not halt the staph infections that kill most burn victims—which prompted the first civilian use of a miracle elixir that was being secretly developed in government-sponsored labs and that ultimately ushered in a new age in therapeutics. Military officials quickly realized this disaster could provide an excellent natural laboratory to test the effectiveness of this drug and see if it could be used to treat the acute traumas of combat in this unfortunate civilian approximation of battlefield conditions. At the time, the very existence of this wondrous medicine—penicillin—was a closely guarded military secret.

From Forgotten Lab Experiment to Wonder Drug

In 1928, Alexander Fleming discovered the curative powers of penicillin, which promised to eradicate infectious pathogens that killed millions every year. But the road to mass producing enough of the highly unstable mold was littered with seemingly unsurmountable obstacles and it remained a forgotten laboratory curiosity for over a decade. But Fleming never gave up and penicillin's eventual rescue from obscurity was a landmark in scientific history.

In 1940, a group at Oxford University, funded in part by the Rockefeller Foundation, isolated enough penicillin to test it on twenty-five mice, which had been infected with lethal doses of streptococci. Its therapeutic effects were miraculous—the untreated mice died within hours, while the treated ones played merrily in their cages, undisturbed. Subsequent tests on a handful of patients, who were brought back from the brink of death, confirmed that penicillin was indeed a wonder drug. But Britain was then being ravaged by the German Luftwaffe during the Blitz, and there were simply no resources to devote to penicillin during the Nazi onslaught.

In June of 1941, two of the Oxford researchers, Howard Florey and Ernst Chain, embarked on a clandestine mission to enlist American aid. Samples of the temperamental mold were stored in their coats. By October, the Roosevelt Administration had recruited four companies—Merck, Squibb, Pfizer and Lederle—to team up in a massive, top-secret development program. Merck, which had more experience with fermentation procedures, swiftly pulled away from the pack and every milligram they produced was zealously hoarded.

After the nightclub fire, the government ordered Merck to dispatch to Boston whatever supplies of penicillin that they could spare and to refine any crude penicillin broth brewing in Merck's fermentation vats. After working in round-the-clock relays over the course of three days, on the evening of December 1st, 1942, a refrigerated truck containing thirty-two liters of injectable penicillin left Merck's Rahway, New Jersey plant. It was accompanied by a convoy of police escorts through four states before arriving in the pre-dawn hours at Massachusetts General Hospital. Dozens of people were rescued from near-certain death in the first public demonstration of the powers of the antibiotic, and the existence of penicillin could no longer be kept secret from inquisitive reporters and an exultant public. The next day, the Boston Globe called it "priceless" and Time magazine dubbed it a "wonder drug."

Within fourteen months, penicillin production escalated exponentially, churning out enough to save the lives of thousands of soldiers, including many from the Normandy invasion. And in October 1945, just weeks after the Japanese surrender ended World War II, Alexander Fleming, Howard Florey and Ernst Chain were awarded the Nobel Prize in medicine. But penicillin didn't just save lives—it helped build some of the most innovative medical and scientific companies in history, including Merck, Pfizer, Glaxo and Sandoz.

"Every war has given us a new medical advance," concludes Marshall. "And penicillin was the great scientific advance of World War II."

Linda Marsa
Linda Marsa is a contributing editor at Discover, a former Los Angeles Times reporter and author of Fevered: Why a Hotter Planet Will Harm Our Health and How We Can Save Ourselves (Rodale, 2013), which the New York Times called “gripping to read.” Her work has been anthologized in The Best American Science Writing, and she has written for numerous publications, including Newsweek, U.S. News & World Report, Nautilus, Men’s Journal, Playboy, Pacific Standard and Aeon.
Abortions Before Fetal Viability Are Legal: Might Science and the Change on the Supreme Court Undermine That?

The United States Supreme Court Building in Washington, D.C.

Unsplash

This article is part of the magazine, "The Future of Science In America: The Election Issue," co-published by LeapsMag, the Aspen Institute Science & Society Program, and GOOD.

Viability—the potential for a fetus to survive outside the womb—is a core dividing line in American law. For almost 50 years, the Supreme Court of the United States has struck down laws that ban all or most abortions, ruling that women's constitutional rights include choosing to end pregnancies before the point of viability. Once viability is reached, however, states have a "compelling interest" in protecting fetal life. At that point, states can choose to ban or significantly restrict later-term abortions provided states allow an exception to preserve the life or health of the mother.

This distinction between a fetus that could survive outside its mother's body, albeit with significant medical intervention, and one that could not, is at the heart of the court's landmark 1973 decision in Roe v. Wade. The framework of viability remains central to the country's abortion law today, even as some states have passed laws in the name of protecting women's health that significantly undermine Roe. Over the last 30 years, the Supreme Court has upheld these laws, which have the effect of restricting pre-viability abortion access, imposing mandatory waiting periods, requiring parental consent for minors, and placing restrictions on abortion providers.

Keep Reading Keep Reading
Josephine Johnston
Josephine Johnston is Director of Research and a Research Scholar at The Hastings Center, an independent bioethics research institute in Garrison, New York. She works on the ethics of emerging biotechnologies, particularly as used in human reproduction, psychiatry, genetics, and neuroscience. Her scholarly work has appeared in medical, scientific, policy, law, and bioethics journals, including New England Journal of Medicine, Science, Nature, Hastings Center Report, and Journal of Law, Medicine and Ethics. She has also written for Stat News, New Republic, Time, Washington Post, and The Scientist, and is frequently interviewed by journalists. Ms. Johnston holds degrees in law and bioethics from the University of Otago in New Zealand. Her current research addresses developments in genetics, including prenatal testing, gene editing, and newborn sequencing.
Democratize the White Coat by Honoring Black, Indigenous, and People of Color in Science

Celebrating BIPOC's achievements in science are a strong first step to make science a guiding force for all.

Unsplash

This article is part of the magazine, "The Future of Science In America: The Election Issue," co-published by LeapsMag, the Aspen Institute Science & Society Program, and GOOD.


Journalists, educators, and curators have responded to Black Lives Matter by highlighting the history and achievements of Black Americans in a variety of fields, including science. The movement has also sparked important demands to address longstanding scientific inequities such as lack of access to quality healthcare and the disproportionate impact of climate change and environmental pollution on neighborhoods of Black, Indigenous, and people of color (BIPOC). Making such improvements requires bringing BIPOC into science and into positions of leadership in laboratories, graduate schools, medical practices, and clinical trials. The moment is right to challenge scientific gatekeepers to respond to Black Lives Matter by widening the pathways that determine who becomes a scientist, a researcher, or a clinician.


The scientific workforce has long lacked diversity, which in turn discourages Black people from pursuing such careers. Causes include a dearth of mentors and role models, preconceived notions that science is exclusive to white males, and subpar STEM education. Across race, gender, class, ability, and all other dimensions that inform how an individual navigates the world, from the familial to the global level, seeing role models who resemble you impacts what you strive for and believe possible. As Marian Wright Edelman stated, "You can't be what you can't see"—a truth with ever-increasing resonance since the U.S. is projected to be minority-white by 2045.

Keep Reading Keep Reading
Garance Choko and Aaron Mertz
Garance Choko started her career as a concert pianist at a very young age. Later, when she moved to the United States to continue her performance studies, she pursued her passion for public administration and innovation. She earned her Masters of Public Administration from Cornell University. Garance has launched innovation firms and designed and implemented physical spaces, national and local health care systems, nationwide public administration processes, and labor policies for institutions, corporations, and governments in North America, Europe, Africa, and the Caribbean. - - - Aaron F. Mertz, Ph.D., is a biophysicist, science advocate, and the founding Director of the Aspen Institute Science & Society Program, launched in 2019 to help foster a diverse scientific workforce whose contributions extend beyond the laboratory and to generate greater public appreciation for science as a vital tool to address global challenges. He completed postdoctoral training in cell biology at Rockefeller University, a doctorate in physics at Yale University, a master’s degree in the history of science at the University of Oxford as a Rhodes Scholar, and a bachelor’s degree in physics at Washington University in St. Louis.