The Truth Wears Off

7
BEK 20150909 THE TRUTH WEARS OFF Is there something wrong with the scientic method? by Jonah Lehrer DECEMBER 13, 2010 On September 18, 2007, a few dozen neuroscientists, psychiatrists, 1 and drug-company executives gathered in a hotel conference room in 2 Brussels to hear some startling news. It had to do with a class of drugs 3 known as atypical or second-generation antipsychotics, which came 4 on the market in the early nin eti es. The drugs, sol d und er brand 5 names such as Abilify, Seroquel, and Zyprexa, had been tested on 6 schizophrenics in several large clinical trials, all of which had demon- 7 strated a dramatic decrease in the subjects’ psychiatric symptoms. 8 As a result, second-generation antipsychotics had become one of the 9 fastest- growing and most protable pharmaceutical classes. By 2001, 10 Eli Lilly’s Zyprexa was generating more revenue than Prozac . It re- 11 mains the company’s top-selling drug. 12 But the data presented at the Brussels meeting made it clear that 13 something strange was happening: the therapeutic power of the drugs 14 appeared to be stead ily waning . A recen t study showed an eect 15 that was less than half of that documented in the rst trials, in the 16 early nineteen-ni netie s. Man y rese arc hers began to argue that the 17 expensive pharmaceuticals weren’t any better than rst-generation 18 antipsyc hotics, whi ch have been in use since the ft ies. “In fact, 19 some times they now look even worse,” John Davis, a professor of 20 psychiatry at the University of Illinois at Chicago, told me. 21 Before the eectiveness of a drug can be conrmed, it must be 22 tested and tested again. Dier ent scien tists in dieren t labs need to 23 repeat the protocols and publish their results. The test of replicabil- 24 ity, as it’s known, is the foundation of modern research. Replicability 25 is how the communit y enforces itself . It’s a safegu ard for the creep 26 of subjectivit y. Most of the time, scientis ts know what resu lts they 27 want, and that can inuen ce the results they get. The premise of 28 replicability is that the scientic community can correct for these  29 aws.  30 But now all sorts of well-established, multiply conrmed ndings  31 hav e started to look increasi ngly uncertain. It’s as if our facts were  32 losing their truth: claims that have been enshrined in textbooks are  33 sudd enly unprov able. This phenome non doesn’t yet have an ocial  34 name, but it’s occurring across a wide range of elds, from psychology  35 to ecology. In the eld of medicine, the phenomenon seems extremely  36 widespread, aecting not only antipsychotics but also therapies rang-  37 ing from cardiac stents to Vitamin E and antidepressants: Davis has  38 a forthcoming analysis demonstrating that the ecacy of antidepres-  39 sants has gone down as much as threefold in recent decades.  40 For many scientists, the eect is especially troubling because of  41 what it exposes about the scien tic proces s. If replica tion is what  42 separ ates the rigor of scien ce from the squis hines s of pseudoscience,  43 where do we put all these rigorously validated ndings that can no  44 longer be proved? Which results should we believe? F rancis Bacon,  45 the early-modern philosopher and pioneer of the scientic method,  46 once declared that experiments were essential, because they allowed  47 us to “put nature to the question.” But it appears that nature often  48 gives us dierent answers.  49 Jonathan Schooler was a young graduate student at the University  50 of Washington in the nineteen-eighties when he discovered a surpris-  51 ing new fact about langu age and memory . At the time, it was widely  52 believed that the act of describing our memories improved them. But,  53 in a series of clever experiments, Schooler demonstrated that subjects  54 shown a face and asked to describe it were much less likely to recog-  55 nize the face when shown it later than those who had simply looked  56 at it. Schooler called the phenomenon “verbal overshadowing.”  57 The study turned him into an academic star. Since its initial pub-  58 lication, in 1990, it has been cited more than four hundred times. Be-  59 fore long, Schooler had extended the model to a variety of other tasks,  60 such as remembering the taste of a wine, identifying the best straw-  61 berry jam, and solving dicult creati ve puzzle s. In each instanc e,  62 asking people to put their perceptions into words led to dramatic  63 1

Transcript of The Truth Wears Off

Page 1: The Truth Wears Off

7/23/2019 The Truth Wears Off

http://slidepdf.com/reader/full/the-truth-wears-off 1/7

BEK 20150909

THE TRUTH WEARS OFFIs there something wrong with the scientific method?

by Jonah Lehrer

DECEMBER 13, 2010

On September 18, 2007, a few dozen neuroscientists, psychiatrists,1

and drug-company executives gathered in a hotel conference room in2

Brussels to hear some startling news. It had to do with a class of drugs3

known as atypical or second-generation antipsychotics, which came4

on the market in the early nineties. The drugs, sold under brand5

names such as Abilify, Seroquel, and Zyprexa, had been tested on6

schizophrenics in several large clinical trials, all of which had demon-7

strated a dramatic decrease in the subjects’ psychiatric symptoms.8

As a result, second-generation antipsychotics had become one of the9

fastest- growing and most profitable pharmaceutical classes. By 2001,10

Eli Lilly’s Zyprexa was generating more revenue than Prozac. It re-11

mains the company’s top-selling drug.12

But the data presented at the Brussels meeting made it clear that13

something strange was happening: the therapeutic power of the drugs14

appeared to be steadily waning. A recent study showed an effect15

that was less than half of that documented in the first trials, in the16

early nineteen-nineties. Many researchers began to argue that the17

expensive pharmaceuticals weren’t any better than first-generation18

antipsychotics, which have been in use since the fifties. “In fact,19

sometimes they now look even worse,” John Davis, a professor of 20

psychiatry at the University of Illinois at Chicago, told me.21

Before the effectiveness of a drug can be confirmed, it must be22

tested and tested again. Different scientists in different labs need to23

repeat the protocols and publish their results. The test of replicabil-24

ity, as it’s known, is the foundation of modern research. Replicability25

is how the community enforces itself. It’s a safeguard for the creep26

of subjectivity. Most of the time, scientists know what results they27

want, and that can influence the results they get. The premise of 28

replicability is that the scientific community can correct for these   29

flaws.   30

But now all sorts of well-established, multiply confirmed findings   31

have started to look increasingly uncertain. It’s as if our facts were   32

losing their truth: claims that have been enshrined in textbooks are   33

suddenly unprovable. This phenomenon doesn’t yet have an official   34

name, but it’s occurring across a wide range of fields, from psychology   35

to ecology. In the field of medicine, the phenomenon seems extremely   36

widespread, affecting not only antipsychotics but also therapies rang-   37

ing from cardiac stents to Vitamin E and antidepressants: Davis has   38

a forthcoming analysis demonstrating that the efficacy of antidepres-   39

sants has gone down as much as threefold in recent decades.   40

For many scientists, the effect is especially troubling because of    41

what it exposes about the scientific process. If replication is what   42

separates the rigor of science from the squishiness of pseudoscience,   43

where do we put all these rigorously validated findings that can no   44

longer be proved? Which results should we believe? Francis Bacon,   45

the early-modern philosopher and pioneer of the scientific method,   46

once declared that experiments were essential, because they allowed   47

us to “put nature to the question.” But it appears that nature often   48

gives us different answers.   49

Jonathan Schooler was a young graduate student at the University   50

of Washington in the nineteen-eighties when he discovered a surpris-   51

ing new fact about language and memory. At the time, it was widely   52

believed that the act of describing our memories improved them. But,   53

in a series of clever experiments, Schooler demonstrated that subjects  54

shown a face and asked to describe it were much less likely to recog-   55

nize the face when shown it later than those who had simply looked   56

at it. Schooler called the phenomenon “verbal overshadowing.”   57

The study turned him into an academic star. Since its initial pub-   58

lication, in 1990, it has been cited more than four hundred times. Be-   59

fore long, Schooler had extended the model to a variety of other tasks,   60

such as remembering the taste of a wine, identifying the best straw-   61

berry jam, and solving difficult creative puzzles. In each instance,   62

asking people to put their perceptions into words led to dramatic   63

1

Page 2: The Truth Wears Off

7/23/2019 The Truth Wears Off

http://slidepdf.com/reader/full/the-truth-wears-off 2/7

BEK 20150909

decreases in performance.64

But while Schooler was publishing these results in highly rep-65

utable journals, a secret worry gnawed at him: it was proving difficult66

to replicate his earlier findings. “I’d often still see an effect, but the67

effect just wouldn’t be as strong,” he told me. “It was as if verbal68

overshadowing, my big new idea, was getting weaker.” At first, he69

assumed that he’d made an error in experimental design or a statis-70

tical miscalculation. But he couldn’t find anything wrong with his71

research. He then concluded that his initial batch of research subjects72

must have been unusually susceptible to verbal overshadowing. (John73

Davis, similarly, has speculated that part of the drop-off in the ef-74

fectiveness of antipsychotics can be attributed to using subjects who75

suffer from milder forms of psychosis which are less likely to show76

dramatic improvement.) “It wasn’t a very satisfying explanation,”77

Schooler says. “One of my mentors told me that my real mistake was78

trying to replicate my work. He told me doing that was just setting79

myself up for disappointment.”80

Schooler tried to put the problem out of his mind; his colleagues81

assured him that such things happened all the time. Over the next82

few years, he found new research questions, got married and had83

kids. But his replication problem kept on getting worse. His first84

attempt at replicating the 1990 study, in 1995, resulted in an effect85

that was thirty per cent smaller. The next year, the size of the effect86

shrank another thirty per cent. When other labs repeated Schooler’s87

experiments, they got a similar spread of data, with a distinct down-88

ward trend. “This was profoundly frustrating,” he says. “It was as89

if nature gave me this great result and then tried to take it back.”90

In private, Schooler began referring to the problem as “cosmic ha-91

bituation,” by analogy to the decrease in response that occurs when92

individuals habituate to particular stimuli. “Habituation is why you93

don’t notice the stuff that’s always there,” Schooler says. “It’s an94

inevitable process of adjustment, a ratcheting down of excitement.95

I started joking that it was like the cosmos was habituating to my96

ideas. I took it very personally.”97

Schooler is now a tenured professor at the University of California98

at Santa Barbara. He has curly black hair, pale-green eyes, and the   99

relaxed demeanor of someone who lives five minutes away from his   100

favorite beach. When he speaks, he tends to get distracted by his   101

own digressions. He might begin with a point about memory, which   102

reminds him of a favorite William James quote, which inspires a   103

long soliloquy on the importance of introspection. Before long, we’re   104

looking at pictures from Burning Man on his iPhone, which leads us   105

back to the fragile nature of memory.   106

Although verbal overshadowing remains a widely accepted the-   107

ory—it’s often invoked in the context of eyewitness testimony, for   108

instance—Schooler is still a little peeved at the cosmos. “I know I   109

should just move on already,” he says. “I really should stop talking   110

about this. But I can’t.” That’s because he is convinced that he has   111

stumbled on a serious problem, one that afflicts many of the most   112

exciting new ideas in psychology.   113

One of the first demonstrations of this mysterious phenomenon   114

came in the early nineteen- thirties. Joseph Banks Rhine, a psy-   115

chologist at Duke, had developed an interest in the possibility of    116

extrasensory perception, or E.S.P. Rhine devised an experiment fea-   117

turing Zener cards, a special deck of twenty-five cards printed with   118

one of five different symbols: a card was drawn from the deck and   119

the subject was asked to guess the symbol. Most of Rhine’s subjects   120

guessed about twenty per cent of the cards correctly, as you’d expect,   121

but an undergraduate named Adam Linzmayer averaged nearly fifty   122

per cent during his initial sessions, and pulled off several uncanny   123

streaks, such as guessing nine cards in a row. The odds of this hap-  124

pening by chance are about one in two million. Linzmayer did it   125

three times.   126

Rhine documented these stunning results in his notebook and pre-   127

pared several papers for publication. But then, just as he began to   128

believe in the possibility of extrasensory perception, the student lost   129

his spooky talent. Between 1931 and 1933, Linzmayer guessed at the   130

identity of another several thousand cards, but his success rate was   131

now barely above chance. Rhine was forced to conclude that the stu-   132

dent’s “extra-sensory perception ability has gone through a marked   133

2

Page 3: The Truth Wears Off

7/23/2019 The Truth Wears Off

http://slidepdf.com/reader/full/the-truth-wears-off 3/7

BEK 20150909

decline.” And Linzmayer wasn’t the only subject to experience such a134

drop-off: in nearly every case in which Rhine and others documented135

E.S.P. the effect dramatically diminished over time. Rhine called this136

trend the “decline effect.”137

Schooler was fascinated by Rhine’s experimental struggles. Here138

was a scientist who had repeatedly documented the decline of his139

data; he seemed to have a talent for finding results that fell apart. In140

2004, Schooler embarked on an ironic imitation of Rhine’s research:141

he tried to replicate this failure to replicate. In homage to Rhine’s142

interests, he decided to test for a parapsychological phenomenon143

known as precognition. The experiment itself was straightforward:144

he flashed a set of images to a subject and asked him or her to identify145

each one. Most of the time, the response was negative—the images146

were displayed too quickly to register. Then Schooler randomly se-147

lected half of the images to be shown again. What he wanted to know148

was whether the images that got a second showing were more likely149

to have been identified the first time around. Could subsequent ex-150

posure have somehow influenced the initial results? Could the effect151

become the cause?152

The craziness of the hypothesis was the p oint: Schooler knows153

that precognition lacks a scientific explanation. But he wasn’t test-154

ing extrasensory powers; he was testing the decline effect. “At first,155

the data looked amazing, just as we’d expected,” Schooler says. “I156

couldn’t believe the amount of precognition we were finding. But157

then, as we kept on running subjects, the effect size”—a standard158

statistical measure—“kept on getting smaller and smaller.” The sci-159

entists eventually tested more than two thousand undergraduates.160

“In the end, our results looked just like Rhine’s,” Schooler said. “We161

found this strong paranormal effect, but it disappeared on us.”162

The most likely explanation for the decline is an obvious one:163

regression to the mean. As the experiment is repeated, that is, an164

early statistical fluke gets cancelled out. The extrasensory powers of 165

Schooler’s subjects didn’t decline—they were simply an illusion that166

vanished over time. And yet Schooler has noticed that many of the167

data sets that end up declining seem statistically solid—that is, they168

contain enough data that any regression to the mean shouldn’t be   169

dramatic. “These are the results that pass all the tests,” he says.   170

“The odds of them being random are typically quite remote, like one   171

in a million. This means that the decline effect should almost never   172

happen. But it happens all the time! Hell, it’s happened to me mul-   173

tiple times.” And this is why Schooler believes that the decline effect   174

deserves more attention: its ubiquity seems to violate the laws of    175

statistics. “Whenever I start talking about this, scientists get very   176

nervous,” he says. “But I still want to know what happened to my   177

results. Like most scientists, I assumed that it would get easier to doc-   178

ument my effect over time. I’d get b etter at doing the experiments,   179

at zeroing in on the conditions that produce verbal overshadowing.   180

So why did the opposite happen? I’m convinced that we can use the   181

tools of science to figure this out. First, though, we have to admit   182

that we’ve got a problem.”   183

In 1991, the Danish zoologist Anders Møller, at Uppsala Univer-   184

sity, in Sweden, made a remarkable discovery about sex, barn swal-   185

lows, and symmetry. It had long been known that the asymmetrical   186

appearance of a creature was directly linked to the amount of muta-   187

tion in its genome, so that more mutations led to more “fluctuating   188

asymmetry.” (An easy way to measure asymmetry in humans is to   189

compare the length of the fingers on each hand.) What Møller discov-   190

ered is that female barn swallows were far more likely to mate with   191

male birds that had long, symmetrical feathers. This suggested that   192

the picky females were using symmetry as a proxy for the quality of    193

male genes. Møller’s paper, which was published in Nature, set off   194

a frenzy of research. Here was an easily measured, widely applicable   195

indicator of genetic quality, and females could be shown to gravitate   196

toward it. Aesthetics was really about genetics.   197

In the three years following, there were ten independent tests of    198

the role of fluctuating asymmetry in sexual selection, and nine of    199

them found a relationship between symmetry and male reproductive   200

success. It didn’t matter if scientists were looking at the hairs on   201

fruit flies or replicating the swallow studies—females seemed to pre-   202

fer males with mirrored halves. Before long, the theory was applied   203

3

Page 4: The Truth Wears Off

7/23/2019 The Truth Wears Off

http://slidepdf.com/reader/full/the-truth-wears-off 4/7

BEK 20150909

to humans. Researchers found, for instance, that women preferred204

the smell of symmetrical men, but only during the fertile phase of 205

the menstrual cycle. Other studies claimed that females had more206

orgasms when their partners were symmetrical, while a paper by an-207

thropologists at Rutgers analyzed forty Jamaican dance routines and208

discovered that symmetrical men were consistently rated as better209

dancers.210

Then the theory started to fall apart. In 1994, there were four-211

teen published tests of symmetry and sexual selection, and only eight212

found a correlation. In 1995, there were eight papers on the subject,213

and only four got a positive result. By 1998, when there were twelve214

additional investigations of fluctuating asymmetry, only a third of 215

them confirmed the theory. Worse still, even the studies that yielded216

some positive result showed a steadily declining effect size. Between217

1992 and 1997, the average effect size shrank by eighty per cent.218

And it’s not just fluctuating asymmetry. In 2001, Michael Jen-219

nions, a biologist at the Australian National University, set out to220

analyze “temporal trends” across a wide range of subjects in ecology221

and evolutionary biology. He looked at hundreds of papers and forty-222

four meta-analyses (that is, statistical syntheses of related studies),223

and discovered a consistent decline effect over time, as many of the224

theories seemed to fade into irrelevance. In fact, even when numerous225

variables were controlled for—Jennions knew, for instance, that the226

same author might publish several critical papers, which could dis-227

tort his analysis—there was still a significant decrease in the validity228

of the hypothesis, often within a year of publication. Jennions ad-229

mits that his findings are troubling, but expresses a reluctance to talk230

about them publicly. “This is a very sensitive issue for scientists,”231

he says. “You know, we’re supposed to be dealing with hard facts,232

the stuff that’s supposed to stand the test of time. But when you see233

these trends you become a little more skeptical of things.”234

What happened? Leigh Simmons, a biologist at the University235

of Western Australia, suggested one explanation when he told me236

about his initial enthusiasm for the theory: “I was really excited237

by fluctuating asymmetry. The early studies made the effect look238

very robust.” He decided to conduct a few experiments of his own,   239

investigating symmetry in male horned beetles. “Unfortunately, I   240

couldn’t find the effect,” he said. “But the worst part was that when   241

I submitted these null results I had difficulty getting them published.   242

The journals only wanted confirming data. It was too exciting an   243

idea to disprove, at least back then.” For Simmons, the steep rise and   244

slow fall of fluctuating asymmetry is a clear example of a scientific   245

paradigm, one of those intellectual fads that both guide and constrain   246

research: after a new paradigm is proposed, the peer-review process   247

is tilted toward positive results. But then, after a few years, the   248

academic incentives shift—the paradigm has become entrenched—so   249

that the most notable results are now those that disprove the theory.   250

Jennions, similarly, argues that the decline effect is largely a prod-   251

uct of publication bias, or the tendency of scientists and scientific   252

 journals to prefer positive data over null results, which is what hap-   253

pens when no effect is found. The bias was first identified by the   254

statistician Theodore Sterling, in 1959, after he noticed that ninety-   255

seven per cent of all published psychological studies with statistically   256

significant data found the effect they were looking for. A “signif-   257

icant” result is defined as any data point that would be produced   258

by chance less than five per cent of the time. This ubiquitous test   259

was invented in 1922 by the English mathematician Ronald Fisher,   260

who picked five per cent as the boundary line, somewhat arbitrar-   261

ily, because it made pencil and slide-rule calculations easier. Sterling   262

saw that if ninety- seven per cent of psychology studies were proving   263

their hypotheses, either psychologists were extraordinarily lucky or  264

they published only the outcomes of successful experiments. In recent   265

years, publication bias has mostly been seen as a problem for clinical   266

trials, since pharmaceutical companies are less interested in publish-   267

ing results that aren’t favorable. But it’s becoming increasingly clear   268

that publication bias also produces major distortions in fields without   269

large corporate incentives, such as psychology and ecology.   270

While publication bias almost certainly plays a role in the decline   271

effect, it remains an incomplete explanation. For one thing, it fails   272

to account for the initial prevalence of positive results among studies   273

4

Page 5: The Truth Wears Off

7/23/2019 The Truth Wears Off

http://slidepdf.com/reader/full/the-truth-wears-off 5/7

BEK 20150909

that never even get submitted to journals. It also fails to explain the274

experience of people like Schooler, who have been unable to replicate275

their initial data despite their best efforts. Richard Palmer, a biologist276

at the University of Alberta, who has studied the problems surround-277

ing fluctuating asymmetry, suspects that an equally significant issue278

is the selective reporting of results—the data that scientists choose to279

document in the first place. Palmer’s most convincing evidence relies280

on a statistical tool known as a funnel graph. When a large number281

of studies have been done on a single subject, the data should follow282

a pattern: studies with a large sample size should all cluster around a283

common value— the true result—whereas those with a smaller sam-284

ple size should exhibit a random scattering, since they’re subject to285

greater sampling error. This pattern gives the graph its name, since286

the distribution resembles a funnel.287

The funnel graph visually captures the distortions of selective re-288

porting. For instance, after Palmer plotted every study of fluctuating289

asymmetry, he noticed that the distribution of results with smaller290

sample sizes wasn’t random at all but instead skewed heavily toward291

positive results. Palmer has since documented a similar problem in292

several other contested subject areas. “Once I realized that selective293

reporting is everywhere in science, I got quite depressed,” Palmer told294

me. “As a researcher, you’re always aware that there might be some295

nonrandom patterns, but I had no idea how widespread it is.” In a296

recent review article, Palmer summarized the impact of selective re-297

porting on his field: “We cannot escape the troubling conclusion that298

some—perhaps many—cherished generalities are at best exaggerated299

in their biological significance and at worst a collective illusion nur-300

tured by strong a-priori beliefs often repeated.”301

Palmer emphasizes that selective reporting is not the same as302

scientific fraud. Rather, the problem seems to be one of subtle omis-303

sions and unconscious misperceptions, as researchers struggle to make304

sense of their results. Stephen Jay Gould referred to this as the “shoe-305

horning” process. “A lot of scientific measurement is really hard,”306

Simmons told me. “If you’re talking about fluctuating asymmetry,307

then it’s a matter of minuscule differences between the right and left308

sides of an animal. It’s millimetres of a tail feather. And so maybe a   309

researcher knows that he’s measuring a good male”—an animal that   310

has successfully mated—“and he knows that it’s supposed to be sym-   311

metrical. Well, that act of measurement is going to be vulnerable to   312

all sorts of perception biases. That’s not a cynical statement. That’s   313

 just the way human beings work.”   314

One of the classic examples of selective reporting concerns the   315

testing of acupuncture in different countries. While acupuncture is   316

widely accepted as a medical treatment in various Asian countries, its   317

use is much more contested in the West. These cultural differences   318

have profoundly influenced the results of clinical trials. Between 1966   319

and 1995, there were forty- seven studies of acupuncture in China,   320

Taiwan, and Japan, and every single trial concluded that acupunc-   321

ture was an effective treatment. During the same period, there were   322

ninety-four clinical trials of acupuncture in the United States, Swe-   323

den, and the U.K., and only fifty-six per cent of these studies found   324

any therapeutic benefits. As Palmer notes, this wide discrepancy sug-   325

gests that scientists find ways to confirm their preferred hypothesis,   326

disregarding what they don’t want to see. Our beliefs are a form of    327

blindness.   328

John Ioannidis, an epidemiologist at Stanford University, argues   329

that such distortions are a serious issue in biomedical research. “These   330

exaggerations are why the decline has become so common,” he says.   331

“It’d be really great if the initial studies gave us an accurate sum-   332

mary of things. But they don’t. And so what happens is we waste   333

a lot of money treating millions of patients and doing lots of follow-  334

up studies on other themes based on results that are misleading.”   335

In 2005, Ioannidis published an article in the Journal of the Amer-   336

ican Medical Association that looked at the forty-nine most cited   337

clinical-research studies in three major medical journals. Forty-five   338

of these studies reported positive results, suggesting that the inter-   339

vention being tested was effective. Because most of these studies   340

were randomized controlled trials—the “gold standard” of medical   341

evidence—they tended to have a significant impact on clinical prac-   342

tice, and led to the spread of treatments such as hormone replacement   343

5

Page 6: The Truth Wears Off

7/23/2019 The Truth Wears Off

http://slidepdf.com/reader/full/the-truth-wears-off 6/7

BEK 20150909

therapy for menopausal women and daily low-dose aspirin to prevent344

heart attacks and strokes. Nevertheless, the data Ioannidis found345

were disturbing: of the thirty-four claims that had been subject to346

replication, forty-one per cent had either been directly contradicted347

or had their effect sizes significantly downgraded.348

The situation is even worse when a subject is fashionable. In349

recent years, for instance, there have been hundreds of studies on350

the various genes that control the differences in disease risk between351

men and women. These findings have included everything from the352

mutations responsible for the increased risk of schizophrenia to the353

genes underlying hypertension. Ioannidis and his colleagues looked at354

four hundred and thirty-two of these claims. They quickly discovered355

that the vast majority had serious flaws. But the most troubling356

fact emerged when he looked at the test of replication: out of four357

hundred and thirty-two claims, only a single one was consistently358

replicable. “This doesn’t mean that none of these claims will turn359

out to be true,” he says. “But, given that most of them were done360

badly, I wouldn’t hold my breath.”361

According to Ioannidis, the main problem is that too many re-362

searchers engage in what he calls “significance chasing,” or finding363

ways to interpret the data so that it passes the statistical test of 364

significance—the ninety-five-per-cent boundary invented by Ronald365

Fisher. “The scientists are so eager to pass this magical test that366

they start playing around with the numbers, trying to find anything367

that seems worthy,” Ioannidis says. In recent years, Ioannidis has be-368

come increasingly blunt about the pervasiveness of the problem. One369

of his most cited papers has a deliberately provocative title: “Why370

Most Published Research Findings Are False.”371

The problem of selective reporting is rooted in a fundamental372

cognitive flaw, which is that we like proving ourselves right and hate373

being wrong. “It feels good to validate a hypothesis,” Ioannidis said.374

“It feels even better when you’ve got a financial interest in the idea375

or your career depends upon it. And that’s why, even after a claim376

has been systematically disproven”—he cites, for instance, the early377

work on hormone replacement therapy, or claims involving various378

vitamins—“you still see some stubborn researchers citing the first   379

few studies that show a strong effect. They really want to believe   380

that it’s true.”   381

That’s why Schooler argues that scientists need to become more   382

rigorous about data collection before they publish. “We’re wasting   383

too much time chasing after bad studies and underpowered experi-   384

ments,” he says. The current “obsession” with replicability distracts   385

from the real problem, which is faulty design. He notes that no-   386

body even tries to replicate most science papers—there are simply   387

too many. (According to Nature, a third of all studies never even   388

get cited, let alone repeated.) “I’ve learned the hard way to be ex-   389

ceedingly careful,” Schooler says. “Every researcher should have to   390

spell out, in advance, how many subjects they’re going to use, and   391

what exactly they’re testing, and what constitutes a sufficient level   392

of proof. We have the tools to be much more transparent about our   393

experiments.”   394

In a forthcoming paper, Schooler recommends the establishment   395

of an open-source database, in which researchers are required to out-   396

line their planned investigations and document all their results. “I   397

think this would provide a huge increase in access to scientific work   398

and give us a much better way to judge the quality of an experiment,”   399

Schooler says. “It would help us finally deal with all these issues that   400

the decline effect is exposing.”   401

Although such reforms would mitigate the dangers of publica-   402

tion bias and selective reporting, they still wouldn’t erase the decline   403

effect. This is largely because scientific research will always be shad-  404

owed by a force that can’t be curbed, only contained: sheer random-   405

ness. Although little research has been done on the experimental   406

dangers of chance and happenstance, the research that exists isn’t   407

encouraging.   408

In the late nineteen-nineties, John Crabbe, a neuroscientist at the   409

Oregon Health and Science University, conducted an experiment that   410

showed how unknowable chance events can skew tests of replicability.   411

He performed a series of experiments on mouse behavior in three   412

different science labs: in Albany, New York; Edmonton, Alberta; and   413

6

Page 7: The Truth Wears Off

7/23/2019 The Truth Wears Off

http://slidepdf.com/reader/full/the-truth-wears-off 7/7

BEK 20150909

Portland, Oregon. Before he conducted the experiments, he tried to414

standardize every variable he could think of. The same strains of 415

mice were used in each lab, shipped on the same day from the same416

supplier. The animals were raised in the same kind of enclosure, with417

the same brand of sawdust bedding. They had been exposed to the418

same amount of incandescent light, were living with the same number419

of littermates, and were fed the exact same type of chow pellets.420

When the mice were handled, it was with the same kind of surgical421

glove, and when they were tested it was on the same equipment, at422

the same time in the morning.423

The premise of this test of replicability, of course, is that each of 424

the labs should have generated the same pattern of results. “If any set425

of experiments should have passed the test, it should have been ours,”426

Crabbe says. “But that’s not the way it turned out.” In one experi-427

ment, Crabbe injected a particular strain of mouse with cocaine. In428

Portland the mice given the drug moved, on average, six hundred cen-429

timetres more than they normally did; in Albany they moved seven430

hundred and one additional centimetres. But in the Edmonton lab431

they moved more than five thousand additional centimetres. Similar432

deviations were observed in a test of anxiety. Furthermore, these in-433

consistencies didn’t follow any detectable pattern. In Portland one434

strain of mouse proved most anxious, while in Albany another strain435

won that distinction.436

The disturbing implication of the Crabbe study is that a lot of 437

extraordinary scientific data are nothing but noise. The hyperactivity438

of those coked-up Edmonton mice wasn’t an interesting new fact—it439

was a meaningless outlier, a by-product of invisible variables we don’t440

understand. The problem, of course, is that such dramatic findings441

are also the most likely to get published in prestigious journals, since442

the data are both statistically significant and entirely unexpected.443

Grants get written, follow-up studies are conducted. The end result444

is a scientific accident that can take years to unravel.445

This suggests that the decline effect is actually a decline of illusion.446

While Karl Popper imagined falsification occurring with a single,447

definitive experiment—Galileo refuted Aristotelian mechanics in an448

afternoon—the process turns out to be much messier than that. Many   449

scientific theories continue to be considered true even after failing nu-   450

merous experimental tests. Verbal overshadowing might exhibit the   451

decline effect, but it remains extensively relied upon within the field.   452

The same holds for any number of phenomena, from the disappear-   453

ing benefits of second-generation antipsychotics to the weak coupling   454

ratio exhibited by decaying neutrons, which appears to have fallen by   455

more than ten standard deviations between 1969 and 2001. Even the   456

law of gravity hasn’t always been perfect at predicting real-world   457

phenomena. (In one test, physicists measuring gravity by means   458

of deep boreholes in the Nevada desert found a two-and-a-half-per-   459

cent discrepancy between the theoretical predictions and the actual   460

data.) Despite these findings, second-generation antipsychotics are   461

still widely prescribed, and our model of the neutron hasn’t changed.   462

The law of gravity remains the same.   463

Such anomalies demonstrate the slipperiness of empiricism. Al-   464

though many scientific ideas generate conflicting results and suffer   465

from falling effect sizes, they continue to get cited in the textbooks   466

and drive standard medical practice. Why? Because these ideas seem   467

true. Because they make sense. Because we can’t bear to let them   468

go. And this is why the decline effect is so troubling. Not because   469

it reveals the human fallibility of science, in which data are tweaked   470

and beliefs shape perceptions. (Such shortcomings aren’t surprising,   471

at least for scientists.) And not because it reveals that many of our   472

most exciting theories are fleeting fads and will soon be rejected.   473

(That idea has been around since Thomas Kuhn.) The decline effect  474

is troubling because it reminds us how difficult it is to prove anything.   475

We like to pretend that our experiments define the truth for us. But   476

that’s often not the case. Just because an idea is true doesn’t mean it   477

can be proved. And just because an idea can be proved doesn’t mean   478

it’s true. When the experiments are done, we still have to choose   479

what to believe.   480

7