Chapter 11: Inferno — Immoral Logic in an Age of Genetic Manipulation

From Films from the Future: The Technology and Morality of Sci-Fi Movies by Andrew Maynard


“If a plague exists, do you know how many

governments would want it and what they’d do

to get it?”

—Sienna Brooks

Decoding Make-Believe

In 1969, the celebrated environmentalist Paul Ehrlich made a stark

prediction. In a meeting held by the British Institute of Biology, he

claimed that, “By the year 2000, the United Kingdom will simply be

a small group of impoverished islands, inhabited by some seventy

million hungry people, of little concern to the other five to seven

billion inhabitants of a sick world.”[^156]

It’s tempting to quip that Ehrlich was predicting the fallout from

Brexit and the UK’s departure from Europe, and his crystal ball

was simply off by a few years. But what kept him up at night, and

motivated the steady stream of dire warnings flowing from him, was

his certainty that human overpopulation would lead to unmitigated

disaster as we shot past the Earth’s carrying capacity.

I left the UK in 2000 to move to the US, and I’m glad to say that, at

the time, the United Kingdom was still some way from becoming

that “small group of impoverished islands.” Yet despite the nation’s

refusal to bow to Ehrlich’s predictions, his writings on population

crashes and control have continued to capture the imaginations of

people over the years, including, I suspect, that of author and the

brains behind the movie Inferno, Dan Brown.

The movie Inferno is based on the book of the same name by

Dan Brown. It’s perhaps not the deepest movie here, but if

you’re willing to crack open the popcorn and suspend disbelief,

it successfully keeps you on the edge of your seat, as any good

mindless thriller should. And it does provide a rather good starting

point for examining the darker side of technological innovation—

biotechnology in particular—when good intentions lead to

seemingly logical, but not necessarily moral, actions.

Inferno revolves around the charismatic scientist and entrepreneur

Bertrand Zobrist (played by Ben Foster). Zobrist is a brilliant

biotechnologist and genetic engineer who’s devoted to saving the

world. But he has a problem. Just like Ehrlich, Zobrist has done the

math, and realized that our worst enemy is ourselves. In his geniuseyes, no matter what we do to cure sickness, improve quality of

life, and enable people to live longer, all we’re doing is pushing the

Earth ever further beyond the point where it can sustain its human

population. And like Ehrlich, he sees a pending future of disease

and famine and death, with people suffering and dying in their

billions, because we cannot control our profligacy.

Zobrist genuinely wants to make the world a better place. But he

cannot shake this vision of apocalyptic disaster. And he cannot

justify using his science for short-term gains, only for it to lead

to long-term devastation. So he makes a terrible decision. To save

humanity from itself, he creates a genetically engineered virus that

will wipe out much of the world’s population—plunging humanity

back into the dark ages, but giving it the opportunity to reset and

build a more sustainable future as a result. And because it seems

that genius entrepreneurs can’t do anything simply, he arranges for

the virus to be elaborately released at a set time in a mysterious

location somewhere in Europe.

The problem is, the authorities are onto him—the authorities in this

case being an entertainingly fictitious manifestation of the World

Health Organization. As the movie starts, Zobrist is being pursued

by WHO agents who chase him to the top of a bell tower in the

Inferno: Immoral Logic in the Age of Genetic Manipulation

I don’t know if Brown and Ehrlich have ever met. I’d like to think

that they’d get on well. Both have a knack for a turn of phrase

that transforms hyperbole into an art form. And both have an

interest in taking drastic action to curb an out-of-control global

human population.

Italian city of Florence where, rather than reveal his secrets, Zobrist

jumps to his death. But in his pocket, he conveniently has a device

that holds the key to where he’s hidden the virus.

This is where Dan Brown brings in his “symbologist” hero, Harvardbased Robert Langdon (Tom Hanks). Langdon, having proven

himself to be rather good at decoding devilishly complex puzzles in

the past, is the ideal person to follow the trail and save the world.

But he quickly finds himself unwittingly wrapped up in a complex

subterfuge where he’s led to believe the WHO are the bad actors,

and it’s up to him and a young doctor, Sienna Brooks (Felicity

Jones), to track down the virus before they get to it.

What follows is a whirlwind of gorgeous locations (Florence, Venice,

Istanbul), misdirection, plot twists, and nail-biting cliffhangers.

We learn that Sienna is, in fact, Zobrist’s lover, and has been using

Langdon to find the virus so she can release it herself. We also learn

that she’s fooled a clandestine global security organization (headed

up by Harry Simms, who’s played perfectly by Irfan Khan) into

helping her, and they set about convincing Langdon he needs to

solve the puzzle while evading the WHO agents.

The movie ends rather dramatically with the virus being contained

just before it’s released. The bad folks meet a sticky end, Langdon

saves the world, and everyone still standing lives happily ever after.

Without doubt, Inferno is an implausible but fun romp. Yet it

does raise a number of serious issues around science, technology,

and the future. Central to these is the question that Paul Ehrlich

and Bertrand Zobrist share in common: Where does the moral

responsibility lie for the future of humanity, and if we could act

now to avoid future suffering—even though the short-term cost may

be hard to stomach—should we? The movie also touches on the

dangers of advanced genetic engineering, and it brings us back to

a continuing theme in this book: powerful entrepreneurs who not

only have the courage of their convictions, but the means to act on

what they believe.

Let’s start, though, with the question of genetically engineering

biological agents, together with the pros and cons of engineering

pathogens to be even more harmful.

Weaponizing the Genome

In 2012, two groups of scientists published parallel papers in the

prestigious journals Science[^157] and Nature[^158] that described, in some

detail, how to genetically engineer an avian influenza virus. What

made the papers stand out was that these scientists succeeded

in making the virus more infectious, and as a result, far deadlier.

The research sparked an intense debate around the ethics of such

studies, and it led to questions about the wisdom of scientists

publishing details of how to make pathogens harmful in a way that

could enable others to replicate their work.

Inferno: Immoral Logic in the Age of Genetic Manipulation

The teams of scientists, led by virologists Ron Fouchier and

Yoshihiro Kawaoka, were interested in the likelihood of a highly

pathogenic flu virus mutating into something that would present a

potentially catastrophic pandemic threat to humans. The unmodified

virus, referred to by the code H5N1, is known to cause sickness

and death in humans, but it isn’t that easy to transmit from person

to person. Thankfully, the virus isn’t readily transmitted by coughs

and sneezes, and this in turn limits its spread quite considerably.

But this doesn’t mean that the virus couldn’t naturally mutate to

the point where it could successfully be transmitted by air. If this

were to occur (and it’s certainly plausible), we could be facing a flu

pandemic of astronomical proportions.

To get a sense of just how serious such a pandemic could be, we

simply need to look back to 1918, when the so-called “Spanish flu”

swept the world.[^159] The outbreak of Spanish flu in the early 1900s

is estimated to have killed around fifty million people, or around 3

percent of the world’s population at the time. If an equally virulent

infectious disease were unleashed on the world today, this would be

equivalent to over 200 million deaths, a mind-numbing number of

people. However, the relative death toll would likely be far higher

today, as modern global transport systems and the high numbers

of people living close to each other in urban areas would likely

substantially increase infection rates.

It’s this sort of scenario that keeps virologists and infectious-disease

epidemiologists awake at night, and for good reason. It’s highly

likely that, one day, we’ll be facing a pandemic of this magnitude.

Viruses mutate and adapt, and the ones that thrive are often those

that can multiply and spread fast. Here, we know that there are

combinations of properties that make viruses especially deadly,

including human pathogenicity, lack of natural resistance in people,

and airborne transmission. There are plenty of viruses that have

one, or possibly two, of these features, yet there are relatively few

that combine all three. But because of the way that evolution and

biology work, it’s only a matter of time before some lucky virus hits

the jackpot, much as we saw back in 1918.

Because of this, it makes sense to do everything we can to be

prepared for the inevitable, including working out which viruses

are likely to mutate into deadly threats (and how) so we can get

our defenses in order before this happens. And this is what drove

Fouchier, Kawaoka, and their teams to start experimenting on H5N1.

H5N1 is a virus that is deadly to humans, but it has yet to evolve

into a form that is readily transmitted by air. What interested

Fouchier and Kawaoka was how likely it was that such a mutation

would appear, and what we could do to combat the evolved virus

if and when this occurs. To begin to answer this question, they

and their teams of scientists intentionally engineered a deadly

new version of H5N1 in the lab, so they could study it. And this

is where the ethical questions began to get tricky. This type of

study is referred to as “gain-of-function” research, as it increases

the functionality and potential deadliness of the virus. Maybe not

surprisingly, quite a few people were unhappy with what was being

done. Questions were asked, for instance, about what would happen

if the new virus was accidentally released. This was not an idle

question, as it turns out, given a series of incidents where infectious

agents ended up being poorly managed in labs.[^160] But it was the

decision to publicly publish the recipe for this gain-of-function

research that really got people worried.

Both Science and Nature ended up publishing the research and the

methods, but only after an intense international debate about the

wisdom of doing so.[^161] However, the decision was, and remains,

controversial. Proponents of the research argue that we need to be

ready for highly pathogenic and transmissible strains of flu before

Concerns like this prompted a group of scientists to release a

Consensus Statement on the Creation of Potential Pathogens in 2014,

calling for greater responsibility in making such research decisions.[^162]

These largely focused on the unintended consequences of wellmeaning research. But there was also a deeper-seated fear here:

What if someone took this research and intentionally weaponized

a pathogen?

This was one of the issues considered by the US National Science

Advisory Board for Biosecurity as it debated drafts of the H5N1

gain-of-function papers in 2011. In a statement released on

December 20, 2011, the NSABB proposed that that the papers

should not be published in their current form, recommending “the

manuscripts not include the methodological and other details that

could enable replication of the experiments by those who would

seek to do harm.”[^163] However, this caused something of a furor at

the time among scientists. The NSABB is an advisory body in the

US and has no real teeth, yet its recommendations drew accusations

of “censorship”[^164] in a scientific community that deeply values

academic freedom.

The NSABB eventually capitulated, and supported the publication

of both papers as they finally appeared in 2012—including the

embedded “how-to” instructions for creating a virulent virus.[^165]

But the question of intentionally harmful use remained. And it’s

concerns like this that underpin the plot in Inferno.

Fouchier, Kawaoka, and their teams showed that it is, in principle,

possible to take a potentially dangerous virus and engineer it into

something even more deadly. To the NSABB and others, this raised

Inferno: Immoral Logic in the Age of Genetic Manipulation

they inevitably arise, and this means having the ability to develop

a stockpile of vaccines. This in turn depends on having a sample of

the virus to be protected against. But this type of research makes

many scientists uneasy, especially given the challenges of preventing

inadvertent releases.

a clear national security issue: What if an enemy nation or a terrorist

group used the research to create a weaponized virus? Echoes of

this discussion stretched back to the 2001 anthrax attacks in the US,

where the idea of “weaponizing” a pathogenic organism became part

of our common language. Since then, discussions over whether and

how biological agents may be weaponized have become increasingly

common.

Intuitively, genetically engineering a virus to weaponize it feels

like it should be a serious threat. It’s easy to imagine the mayhem

a terrorist group could create by unleashing an enhanced form

of smallpox, Ebola, or even the flu. Thankfully, most biosecurity

experts believe that the risks are low here. Despite these imagined

scenarios, it takes substantial expertise and specialized facilities to

engineer a weaponized pathogen, and even then, it’s unclear that

the current state of science is good enough to create an effective

weapon of terror. More than this, though, most experts agree that

there are far easier and cheaper ways of creating terror, or taking

out enemy forces, than using advanced biology. And because of

this, it’s hard to find compelling reasons why an organization would

weaponize a pathogen, rather than using far easier and cheaper

ways of causing harm. Why spend millions of dollars and years

of research on something that may not work, when you can do

more damage with less effort using a cell phone and home-made

explosives, or even a rental truck? The economics of weaponized

viruses simply don’t work outside of science fiction thrillers and

blockbuster movies. At least, not in a conventional sense.

And this is where Inferno gets interesting, as Zobrist is not terrorist

in the conventional sense. Zobrist’s aim is not to bring about change

through terror, but to be the agent of change. And his mechanism

of choice is a gain-of-function genetically engineered virus. Unlike

the potential use of genetically modified pathogens by terrorists,

or even nation-states, the economics of Zobrist’s decision actually

make some sense, warped as they are. In his mind, he envisions

a cataclysmic future for humanity, brought about through outof-control overpopulation. and he sees it as a moral imperative

to use his expertise and wealth to help avoid it, albeit by rather

drastic means.

As this is movie make-believe, the technology Zobrist ends up

developing is rather implausible. But it’s not that far-fetched.

Certainly, we know from the work of Fouchier, Kawaoka, and

others that it is possible to engineer viruses to be more deadly

Immoral Logic?

Some years ago, my wife gave me a copy of Daniel Quinn’s book

Ishmael. The novel, which won the Turner Tomorrow Award in

1991, has something of a cult following. But I must confess I was

rather disturbed by the arguments it promoted. What concerned me

most, perhaps, was a seemingly pervasive logic through the book

that seemed to depend on “ends,” as defined by a single person,

justifying extreme “means” to get there. Echoing both Paul Ehrlich

and Dan Brown, Quinn was playing with the idea that seemingly

unethical acts in the short term are worth it for long-term prosperity

and well being, especially when, over time, the number of people

benefitting from a decision far outnumber those who suffered as a

consequence.

Ishmael is a Socratic dialogue between the “pupil”—the narrator—

and his “teacher,” a gorilla that has the power of speech and reason.

The book uses this narrative device to dissect human history and

the alleged rise of tendencies that have led to a global culture of

selfish greed, unsustainable waste, and out-of-control population

growth. The book is designed to get the reader to think and reflect.

In doing so, it questions our rights as humans above those of other

organisms, and our obligations to other humans above that to

the future of the Earth as a whole. Many of the underlying ideas

in the book are relatively common in environmentalist thinking.

What Ishmael begins to illuminate, though, is what happens when

some of these ideas are taken to their logical conclusions. One of

those conclusions is that, if the consequence of a growing human

population and indiscriminate abuse of the environment is a

sick and dying planet, anything we do now to curb our excesses

is justified by the future well-being of the Earth and its many

ecosystems. The analogy used by Quinn is that of a surgeon cutting

out a malignant cancer to save the patient, except that, in this case,

Inferno: Immoral Logic in the Age of Genetic Manipulation

than their naturally-occurring counterparts. And we’re not that

far from hypothetically being able to precisely design a virus

with a specific set of characteristics, an ability that will only

accelerate as we increasingly use cyber-based technologies and

artificial-intelligence-based methods in genetic design. Because of

these converging trends in capabilities, when you strip away the

hyperbolic narrative and cliffhanger scenarios from Inferno, there’s

a kernel of plausibility buried in the movie that should probably

worry us, especially in a world where powerful individuals are able

to translate their moral certitude into decisive action.

the patient is the planet, and humanity is both the cancer and the

surgeon.

This is a similar philosophy, of taking radical action in the present

to save the future, that Ehrlich promoted in his 1968 book, The

Population Bomb.[^166] As a scientist and environmentalist, Ehrlich

was appalled by where he saw the future of humanity and Planet

Earth heading. As the human population increased exponentially,

he believed that, left unchecked, people would soon exceed the

carrying capacity of the planet. If this happened, he believed we

would be plunged into a catastrophic cycle of famine, disease, and

death, that would be far worse than any preventative actions we

might take.

Ehrlich opens his book with a dramatic account of him personally

experiencing localized overpopulation in Delhi. This experience

impressed on him that, if this level of compressed humanity was

to spread across the globe (as he believed it would), we would

be responsible for making a living hell for future generations,

something he saw as his moral duty to do what he could to prevent.

In the book, Ehrlich goes on to explore ways in which policies

could be established to avoid what he saw as an impending disaster.

He also looked at ways in which people might be persuaded to

change their habits and beliefs in an attempt to dramatically curb

population growth. But he considered the threat too large to stop

at political action and persuasion. To him, if these failed, drastic

measures were necessary. He lamented, for instance, that India had

not implemented a controversial sterilization program for men as

a means of population control. And he talked of triaging countries

needing aid to avoid famine and disease, by helping only those

that could realistically pull themselves around while not wasting

resources on “hopeless cases.”

Ehrlich’s predictions and views were both extreme and challenging.

And in turn, they were challenged by others. Many of his predictions

have not come to pass, and since publication of The Population

Bomb, Ehrlich has pulled back from some of his more extreme

proposals. There are many, though, who believe that the sheer

horror of his predictions and his proposed remedies scared a

generation into taking action before it was too late. Even so, we are

still left with a philosophy which, much like the one espoused in

Ishmael, suggests that one person’s prediction of pending death and

It is precisely this philosophy that Dan Brown explores through the

character of Zobrist in Inferno. Superficially, Zobrist’s arguments

seem to make sense. Using an exponential growth model of global

population, he predicts a near future where there is a catastrophic

failure of everything we’ve created to support our affluent twentyfirst-century lifestyle. Following his arguments, it’s not hard to

imagine a future where food and water become increasingly

scarce, where power systems fail, leaving people to the mercy of

the elements, where failing access to healthcare leads to rampant

disease, and where people are dying in the streets because they are

starving, sick, and have no hope of rescue.

As well as being a starkly sobering vision, this is also a plausible

one—up to a point. We know that when animal populations get

out of balance, they often crash. And research on complex systems

indicates that the more complex, interdependent, and resourceconstrained a system gets, the more vulnerable it can become to

catastrophic failure. It follows that, as we live increasingly at the

limits of the resources we need to sustain nearly eight billion people

across the planet, it’s not too much of a stretch to imagine that

we are building a society that is very vulnerable indeed to failing

catastrophically. But if this is the case, what do we do about it?

Early on in Inferno, Zobrist poses a question: “There’s a switch.

If you throw it, half the people on earth will die, but if you don’t,

in a hundred years, then the human race will be extinct.” It’s an

extreme formulation of the ideas of Quinn and Ehrlich, and not

unlike a scaled-up version of the Trolley Problem that philosophers

of artificial intelligence and self-driving cars love to grapple with.

But it gets to the essence of the issue at hand: Is it better to kill

a few people now and save many in the future, or to do nothing,

condemning billions to a horrible death, and potentially signing off

on the human race?

Ehrlich and Quinn suggest that it’s moral cowardice to take the “not

my problem” approach to this question. In Inferno, though, Brown

elevates the question from one of philosophical morality to practical

reality. He gives the character of Zobrist the ability to follow through

on his convictions, and to get out of his philosophical armchair to

Inferno: Immoral Logic in the Age of Genetic Manipulation

destruction has greater moral weight than the lives of the people

they are willing to sacrifice to save future generations.

quite literally throw the switch, believing he is saving humanity as

he does so.

The trouble is, this whole scenario, while easy to spin into a

web of seeming rationality, is deeply flawed. Its flaws lie in the

same conceits we see in calls for action based on technological

prediction. It assumes that the future can be predicted from the

exponential trends of the past (a misconception that was addressed

in chapter nine and Transcendence), and it amplifies, rather than

moderates, biases in human reasoning and perception. Reasoning

like this creates an artificial certainty around the highly uncertain

outcomes of what we do, and it justifies actions that are driven by

ideology rather than social responsibility. It also assumes that the

“enlightened,” whoever they are, have the moral right to act, without

consent, on behalf of the “unenlightened.”

In the cold light of day, what you end up with by following such

reasoning is something that looks more like religious terrorism, or

the warped actions of the Unabomber Ted Kaczynski, than a plan

designed to create social good.

This is not to say we are not facing tough issues here. Both the

Earth’s human population and our demands on its finite resources

are increasing in an unsustainable way. And this is leading to serious

challenges that should, under no circumstances, be trivialized.

Yet, as a species, we are also finding ways to adapt and survive,

and to overcome what were previously thought to be immovable

barriers to what could be achieved. In reality, we are constantly

moving the goalposts of what is possible through human ingenuity.

The scientific and social understanding of the 1960s was utterly

inadequate for predicting how global science and society would

develop over the following decades, and as a result, Ehrlich and

others badly miscalculated both the consequences of what they

saw occurring and the measures needed to address them. These

developments included advances in artificial fertilizers and plant

breeding that transformed the ability of agriculture to support a

growing population. We continue to make strides in developing

and using technology to enable a growing number of people to live

sustainably on Earth, so much so that we simply don’t know what

the upper limit of the planet’s sustainable human population might

be. In fact, perhaps the bigger challenge today is not providing

people with enough food, water, and energy, but in overcoming

social and ideological barriers to implementing technologies in ways

that benefit this growing population.

Yet while such thinking can lead to what I believe is an immoral

logic, we cannot afford to dismiss the possibility that inaction in the

present may lead to catastrophic failures in the future. If we don’t

get our various acts together, there’s still a chance that a growing

population, a changing climate, and human greed will lead to

future suffering and death. As we develop increasingly sophisticated

technologies, these only add to the uncertainty of what lies around

the corner. But if we’re going to eschew following an immoral logic,

how do we begin to grapple with these challenges?

The Honest Broker

Perhaps one of the most difficult challenges scientists (and

academics more broadly) face is knowing when to step out of the

lab (or office) and into the messy world of politics, advocacy, and

activism. The trouble is, we’re taught to question assumptions, to be

objective, and to see issues from multiple perspectives. As a result,

many scientists see themselves as seekers of truth, but skeptical

of the truth. Because of this, many of us are uneasy about using

our work to make definitive statements about what people should

or should not be doing. To be quite frank, it feels disingenuous to

set out to convince people to act as if we know the answers to a

problem, when in reality all we know is the limits of our ignorance.

There’s something else though, that makes many scientists leery

about giving advice, and that’s the fear of losing the trust and

respect of others. Many of us have an almost pathological paranoia

of being caught out in an apparent lie if we make definitive

statements in public, and for good reason; there are few problems

in today’s society that have cut-and-dried solutions, and to claim

that there are smacks of charlatanism. More than this, though,

there’s a sense within the culture of science that making definitive

statements in public is more about personal ego than professional

responsibility.

Inferno: Immoral Logic in the Age of Genetic Manipulation

Imagine now that, in 1968, a real-life Zobrist had decided to act

on Ehrlich’s dire predictions and indiscriminately rob people of

their dignity, autonomy, and lives, believing that history would

vindicate them. It would have been a morally abhorrent tragedy

of monumental proportion. This is part of the danger of confusing

exponential predictions with reality, and mixing them up with

ideologies that adhere religiously to a narrow vision of the future, to

the point that its believers are willing to kill for the alleged longterm good of society.

The unwritten rule here sometimes seems to be that scientists

should stick to what they’re good at—asking interesting questions

and discovering interesting things—and leave it to others to decide

what this means for society more broadly. This is, I admit, something

of an exaggeration. But it does capture a tension that many scientists

grapple with as they try to reconcile their primary mission to

generate new knowledge with their responsibility as a human being

to help people not make a complete and utter mess of their lives.

Not surprisingly, these lines become blurred in areas where research

is driven by social concerns. As a result, there’s a strong tradition

in areas like public health of research being used to advocate for

socially beneficial behaviors and policies. And scientists focusing on

environmental sustainability and climate change are often working

in these areas precisely because they want to make a difference.

To many of them, their research isn’t worth their time if it doesn’t

translate into social impact, and that brings with it a responsibility to

advocate for change.

This is the domain that scientists like Paul Ehrlich and Dan Brown’s

Zobrist inhabit. They are engaged in their science because they

see social and environmental problems that need to be solved.

To many researchers in this position, their science is a means to a

bigger end, rather than being an end in itself. In fact, I suspect that

many researchers in these areas of study would argue that there is

a particular type of immorality associated with scientists who, with

their unique perspective, can see an impending disaster coming, and

decide to do nothing about it.

Here, the ethics of the scientist-advocate begin make a lot of sense.

Take this thought experiment, for instance. Imagine your research

involves predicting volcanic eruptions (just to make a change from

population explosions and genetically engineered viruses), and

your models strongly indicate that the supervolcano that lies under

Yellowstone National Park could erupt sometime in the next decade.

What should you do? Do nothing, and you potentially condemn

millions of people—maybe more—to famine, poverty, disease, and

death. Instinctively, this feels like the wrong choice, and I suspect

that few scientists would just ignore the issue. But they might say

that, because of the uncertainty in their predictions, more research is

needed, including more research funding, and maybe a conference

or two to develop the science more and argue over the results. In

other words, there’d probably be lots of activity, but very little action

To some scientists, however, this would be ethically untenable, and

an abdication of responsibility. To them, the ethical option would be

to take positive action: Raise awareness, shock people into taking

the risk seriously, hit the headlines, give TED talks, make people sit

up and listen and care, and, above all, motivate policy makers to do

something. Because—so the thinking would go—even if the chances

are only one in a thousand of the eruption happening, it’s better to

raise the alarm and be wrong than stay silent and be right.

This gets to the heart of the ethics of science-activism. It’s what lies

behind the work of Paul Ehrlich and others, and it’s what motivates

movements and organizations that push for social, political, and

environmental change to protect the future of the planet and its

inhabitants. And yet, compelling as the calculus of saved future

lives is, there is a problem. Pushing for action based on available

evidence always comes with consequences. Sadly, there’s no free

pass if you make a mistake, or the odds don’t fall in your favor.

Going back to the Yellowstone example, a major eruption could

well render large swaths of the mid-US uninhabitable. Agriculture

would be hit hard, with air pollution and localized climate shifts

making living conditions precarious for tens of millions of people.

On the other hand, preparing for a potential eruption would most

likely involve displacing millions of people, possibly leading to

coastal overcrowding, loss of jobs, homelessness, and a deep

economic recession. The outcomes of the precautionary actions—

irrespective of whether the predictions came true or not—would be

devastating for some. They may be seen as worth it in the long run

if the eruption takes place. But if it doesn’t, the decision to act will

have caused far more harm than inaction would have. Now imagine

having the burden of this on your shoulders, because you had the

courage of your scientific convictions, even though you were wrong,

and it becomes clearer why it takes a very brave scientist indeed to

act on the potential consequences of their work.

This is, obviously, an extreme and somewhat contrived example.

But it gets to the core of the dilemma surrounding individuals

acting on their science, and it underlies the tremendous social

responsibility that comes with advocating for change based on

scientific convictions. To make matters worse, while we all like to

think we are rational beings—scientists especially—we are not. We

are all at the mercy of our biases and beliefs, and all too often we

Inferno: Immoral Logic in the Age of Genetic Manipulation

that would help those people who would be affected if such an

eruption did occur.

interpret our science through the lens of these. And this means that

when an individual, no matter how smart they are, decides that they

have compelling evidence that demands costly and disruptive action,

there’s a reasonably good chance that they’ve missed something.

So how do we get out of this bind, where conscientious scientists

seem to be damned if they do, and damned if they don’t? The

one point of reasonable certainty here is that it’s dangerous for

an individual to push an agenda for change on their own. It’s just

too easy for someone to be blinded by what they believe is right

and true, and as a result miss ways forward that are more socially

responsible. At the same time, it’s irresponsible to suggest that

scientists should be seen and not heard, especially when they have

valuable insights into emerging risks and ways to avoid them.

One way forward is in collective advocacy. There’s a much greater

chance of a hundred scientists having a clear view of emerging

challenges and options than one lone genius. And in reality, this is

how science gets translated into action on many large issues. But

this does mean that experts need to be prepared to work together,

and to have the humility to accept that their personal ideas may

need to be reined in or modified for the common good. This is

where most experts are at with big issues like climate change and

vaccines. But there are many other socially important issues that

either don’t rise to the level of collective efforts from scientists, or

are still uncertain enough that there is not enough evidence for a

consensus to emerge. So, what are socially responsible scientists to

do in these cases?

In 2007, the scholar Roger Pielke Jr. grappled with some of these

challenges in his book The Honest Broker: Making Sense of Science

in Policy and Politics.[^167] Pielke was especially interested in how

science and scientists inform policy and operate within the political

arena. Because of this, his book takes quite a narrow view of

advocacy, particularly when it comes to exploring how scientists can

use policy advocacy to bring about change. But much of his analysis

is relevant to any scientist trying to thread the needle of remaining

true to their profession while acting as a responsible citizen.

Pielke astutely recognizes that there is no single best way that

scientists can translate what they know and what they believe to be

true into societally relevant action. Instead, taking his own advice,

he suggests that there are a range of possible options here, with

Pielke characterizes the Pure Scientist as someone simply interested

in generating new knowledge and placing it into a common

reservoir of information, which they leave to others to dip into and

use. In other words, they create a wall between themselves and

the society they live in, assuming that someone else may one day

find some use for what they do. If this sounds a little unrealistic,

it probably is. Even Pielke acknowledges that such scientists are

probably found more frequently in myth than in reality. Yet this

is a relatively common stereotype of scientists, certainly within

Western culture.

Pielke’s next category is the Science Arbiter. This, I suspect,

is where many scientists are the most comfortable. In Pielke’s

framework, Science Arbiters recognize that effective and socially

relevant decisions are made on good evidence and clear information

about the pros and cons of different options. Rather than having

an opinion on what is the right or the wrong decision, Science

Arbiters help ensure people have access to the science and

evidence they need to make the best possible decisions. There is

a twist here, though. Pielke also argues that, because people who

feel comfortable in this role have a deep belief in the scientific

process, they tend to focus on issues that they believe can be

resolved through science, while staying away from those that they

believe cannot.

Then there are scientists—for instance, those working in areas

driven by real-world challenges like health and sustainability—who

feel they cannot morally justify providing what seem to them to be

scientifically sound but socially hollow options to decision makers.

These, in Pielke’s terminology, are the Issue Advocates. They are

scientists on a mission to change the world, to fix what they see as

(mainly) social problems, and to use their science to the best of their

ability to do this. These are people who use science as a means to

an end, and are driven by their own beliefs and convictions. Zobrist

Inferno: Immoral Logic in the Age of Genetic Manipulation

four in particular standing out. These he refers to as four idealized

roles of science in policy and politics, but they apply equally well

to scientists trying to bring about what they consider to be positive

social change. The first of these roles is the Pure Scientist. This

is perhaps closest to the picture of the scientists I drew at the

beginning of this section, the person committed to objectivity and

evidence, who is seriously worried by the idea of making decisions

where there is only uncertainty.

would be considered by Pielke to be an Issue Advocate, as would, I

suspect, Paul Ehrlich.

And finally, there is the Honest Broker. This, in Pielke’s language, is

the person who actively engages with decision-makers to help them

see how science and evidence support (or don’t) the various options

that are open to them. This is the scientist who believes, more than

anything, in helping people make the best decision they can based

on the evidence, but who understands that, ultimately, they don’t

have the right to dictate which decision is made.

Pielke tries not to stand in judgment of the four ways he describes

scientists engaging with politics and policy. But it’s clear from his

writing that he’s a fan of the honest broker. And, to be honest, so

am I. This is the role I try to carve out for myself in my public-facing

work, trying not to judge others or advocate for a specific course

of action, but to help people make the best-informed decisions for

themselves and their communities, based on available evidence

and insights.

This is an approach that, to me, avoids mistaking personal values

for the “right” values, and respects deeply held beliefs and values

in others, even where you may disagree with them. It’s a path

toward empowering others while trying not to let your ego get in

the way. And with most of the issues I grapple with in my work, I’m

comfortable with it, because in most cases there are not bright-line

right or wrong answers.

This Honest Broker role extends to any situation where someone

with useful knowledge and insights is prepared to engage with

people who might benefit from them. Of course, sometimes people

will make decisions that lead to harm anyway. But how much more

tragic if these decisions are made simply because they were never

aware of the alternatives or the consequences. Yet, I’ll be the first

to admit that this role, while being rooted deeply in values that I

consider important, has its problems. And nowhere are they more

apparent than when issues of such moral peril arise that not to

advocate for a certain stance, or a particular way forward, ends up

becoming tacit support for not taking action.

To many, inaction on climate change and the use and proliferation

of nuclear weapons falls into this category, as does the rejection of

vaccines. These are issues where indecision or lack of advocacy has

a high chance of adversely impacting millions of people. In cases

like these, there is increasing pressure to shift from being an Honest

This is, of course, another sticky point, because as soon as an issue

becomes a focus of attention, the battles begin for whose “science”

is the most legitimate. As someone with leanings toward being an

Honest Broker, I would suggest that, where there is uncertainty

in the science (which is pretty much always—that’s the nature of

science), the weight of scientific evidence becomes critical. There are

always going to be multiple ways that science can be interpreted,

but some of these will most likely be more strongly supported by

the evidence than others. And here, nothing good ever comes from

simply selecting the science that supports your issue and rejecting

the science that doesn’t. This is a path to self-delusion, because,

at the end of the day, wishing something is true simply because it

supports what you believe doesn’t make it so.

But then, what do you do if the evidence seems to point toward

a looming catastrophe, and no one’s listening? This is where

charismatic voices like Paul Ehrlich’s arise. And it’s where, as a

society, we need to decide how to respond to what they preach.

Dictating the Future

In the case of Inferno, overpopulation is perceived as a looming

catastrophe that will result in misery and death for hundreds of

millions of people, unless radical action is taken. Zobrist sees this

and believes he has a solution. But, having been effectively outcast

by the scientific community for his radical ideas, he resorts to

drastic measures.

In the movie, Zobrist’s plan to cull half of the world’s population

through his genetically engineered virus is, of course, abhorrent.

This is what provides the dramatic tension that keeps us glued to

the screen, fueled by our moral outrage. But there’s an interesting

Inferno: Immoral Logic in the Age of Genetic Manipulation

Broker to an Issue Advocate. And yet, because of the dangers of

values and belief-driven short-sightedness, even in these cases, it’s

hard to justify one person being the sole arbiter of truth. Rather,

as Pielke argues, this is where we need institutions and sociallysanctioned organizations to act as the instruments of advocacy.

Pielke mentions groups like the National Academy of Sciences, and

by inference, similar organizations around the world. But I suspect

others would include advocacy groups here as well that are focused

on specific issues, yet recognize the importance of science in

advocating for action.

twist here, and it comes not from the movie, but the book that the

film’s based on.

Dan Brown’s book Inferno, like the movie, follows a crazy

countdown as Robert Langdon struggles to unravel the clues left by

Zobrist to the location of the virus. As in the movie, Zobrist believes

enough in the legitimacy of his actions that he’s willing to die rather

than give up his secrets. But then, as the location of the virus is

discovered, the book and the movie diverge quite dramatically.

In the book, Langdon and the WHO arrive too late. The virus has

been released, and has been infecting people for some time. But

surprisingly, no one is dying. It turns out that book-Zobrist didn’t

create a killer virus. Instead, he created a virus that rendered every

third person it infected sterile. What’s more, he ensured that this

“every third person” trait was heritable, meaning that, in every

subsequent generation, one in three people would also be sterile.

In the book, no one died as a result of Zobrist’s genetically modified

virus. Rather, he set in motion a chain of events that would

eventually lead to the Earth’s human population being reduced

to a manageable size. Instead of being the evil scientist intent on

murdering people, he emerges as a lone-genius savior of the future

of humankind.

This outcome intrigues me, as it supports the idea of the lone

visionary scientist as someone who can save the world. And it

suggests that they could probably do it better than a committee of

scientists, because they have a clarity of vision and purpose that a

large and unwieldy group would lack.

I’m pretty sure that the book version of Zobrist’s plan would have

had a profound and ultimately positive impact on the Earth’s human

population. It may also have led to an improved quality of life for

many people, although, humans being humans, there’s also the

chance of self-interest and ignorance putting paid to this possibility.

Yet despite its superficial elegance, something worries me about

the idea of imposing sterility on a third of the world’s population

in the name of social good, and this is the lack of choice that

Zobrist’s victims had. For sure, he “saved” society in the book. But

in doing so, did he end up betraying the individuals that make up

that society?

This is a particularly knotty and ultimately unresolvable moral

question, as it comes down to weighing the good of the many

We’re also still left with the problem that, no matter how much we

delude ourselves, we cannot predict the future. Which means that,

compelling as book-Zobrist’s case was, he had no way of knowing

whether he needlessly condemned a third of the world’s population

to sterility. This was a gamble he was willing to take. But what gave

him the right to take this gamble in the first place? Not the people

whose futures he was playing with, that’s for sure. And this is

ultimately where the challenge lies when it comes to lone scientistadvocates and genius-activists. No matter how compelling their

vision of the future, or how persuasive their solutions to making it

better, where do they get the right to act unilaterally on issues that

ultimately impact us all?

Some, I suspect, would argue that time and necessity are on their

side. I would counter that these are not excuses for preventing

people who are likely to be affected by major decisions from having

a say in their collective future. This, though, means that we need

better ways of making collective decisions as a society (as was seen

in chapter ten and The Man in the White Suit), especially where

technological innovation is both pushing us toward potentially

catastrophic futures and yet is potentially part of the solution to

avoiding such futures. And we need to get better at making such

collective decisions fast, because if there’s one thing that these lone

scientist-advocates have right in many cases, it’s that time is short!

And nowhere is this more apparent than with an issue that’s tightly

coupled to a burgeoning human population: climate change.

Inferno: Immoral Logic in the Age of Genetic Manipulation

against the good of the few. The book version of Zobrist violates

basic human rights by dictating the fate of people infected by his

virus. And I doubt that this would have been a bloodless violation;

while indiscriminate sterilization may seem a small price to pay for

averting world hunger, try telling that to someone desperate for

children who has been robbed of the opportunity, or someone who

depends on growing a family to sustain their livelihood.

[^156]: Bernard Dixon (1971) “In Praise of Prophets.” New Scientist, 16 September 1971, page 606.

[^157]: Sander Herfst and colleagues (2012) “Airborne Transmission of Influenza A/H5N1 Virus Between Ferrets” Science, 336 (6088) pp 1534-1541 http://doi.org/10.1126/science.1213362

[^158]: Masaki Imai and colleagues (2012) “Experimental adaptation of an influenza H5 HA confers respiratory droplet transmission to a reassortant H5 HA/H1N1 virus in ferrets” Nature 486, pp 420–428 http://doi.org/10.1038/nature10831

[^159]: Jeffery K. Taubenberger and David M. Morens (2006) “1918 Influenza: the Mother of All Pandemics“. Emerging Infectious Diseases volume 12, number 1, pages 15-22 https://doi.org/10.3201/eid1201.050979

[^160]: Jocelyn Kaiser (2014) “Lab incidents lead to safety crackdown at CDC.” Published in Science Magazine, July 11, 2014. http://www.sciencemag.org/news/2014/07/lab-incidents-lead-safety-crackdown-cdc

[^161]: Ed Yong (2012) “The risks and benefits of publishing mutant flu studies.” Nature News, March 2, 2012 http://doi.org/10.1038/nature.2012.10138

[^162]: Cambridge Working Group Consensus Statement on the Creation of Potential Pandemic Pathogens (PPPs). http://www.cambridgeworkinggroup.org/

[^163]: Press Statement on the NSABB Review of H5N1 Research, December 20, 2011. https://web.archive.org/web/20160407031930/https://www.nih.gov/news-events/news-releases/press-statement-nsabb-review-h5n1-research

[^164]: Heidi Ledford (2012) “Call to censor flu studies draws fire.” Published in Nature News January 3, 2012. http://doi.org/10.1038/481009a

[^165]: March 29-30, 2012 Meeting of the National Science Advisory Board for Biosecurity to Review Revised Manuscripts on Transmissibility of A/H5N1 Influenza Virus. Statement of the NSABB: https://web.archive.org/web/20190214205704/http://www.virology.ws/NSABB_statement_march_2012.pdf

[^166]: Ehrlich, P. (1968). “The Population Bomb.” Sierra Club/Ballantine Books.

[^167]: Roger A. Pielke Jr. (2007). “The Honest Broker: Making Sense of Science in Policy and Politics” Published by Cambridge University Press.