Science on Trial

When you ask scientists about their work, they cite their positive results. When you ask the public what makes science credible, they cite peer review. When you ask ChatGPT for its sources, it points to publications as if publication were truth. If that all sounds fine, your notion of science is broken.

Science is supposed to be the act of changing beliefs in response to evidence. Every claim stands trial forever. Science is also, now, a noun. A body of credentialed outputs. An institution. The institution was built to support the activity, and has become more rewarding to maintain than the activity itself.

Science today resembles a music rehearsal at Berklee more than it does a debugging session on reality.

The protocol

Four centuries of arguing about what counts as knowledge produced a protocol. Each step exists because someone proved it was necessary.

Each step produces a public artifact: a registered prediction, an alternatives list, a timestamped log, a complete dataset. The trail is what counts as knowledge.

Four terms

Routinely conflated. The exhibits below turn on keeping them apart.

The institution trains the reader to read the first as the fourth. It isn’t.

Now the exhibits.

Exhibit 1. Pre-registration is missing

ClinicalTrials.gov came online in 2000, and prospective registration entered the norm for NHLBI-funded cardiovascular trials around the same time. Kaplan and Irvin (2015) compared 55 large NHLBI trials before and after. Pre-2000: 17 of 30 (57%) reported positive primary outcomes. Post-2000: 2 of 25 (8%). Same question, same funder, chain of custody added.

The mechanism check is sharper than the headline. Of the 25 post-2000 trials, 12 still reported significant secondary outcomes. Without the mandate to pre-declare which outcome counted, 48% would have told a positive story anyway.

Scheel, Schijen, and Lakens (2021) found the same pathology in psychology under a different comparison: standard reports against Registered Reports in the same field. Standard papers: 96% positive. Registered Reports: 44%. The baseline has been this high since Sterling reported 97% in 1959.

In that slice of high-stakes cardiovascular RCTs, the apparent success rate collapsed as soon as outcome declaration became visible. Those findings shaped treatment decisions for millions. Hormone replacement therapy was prescribed to tens of millions of women on the strength of observational signals. The Women’s Health Initiative RCT (2002) was stopped early when its data safety monitoring board found combination HRT raised breast cancer and stroke risk. Prescriptions for combination HRT fell roughly two-thirds within a year.

Exhibit 2. Red-team isn’t happening

Under honest methodology, the published positive-result rate is capped by arithmetic. At α = 0.05 and 80% statistical power:

If the tested hypotheses are…ExpectationReported (Scheel 2021, psychology)
20% true20% positive96% positive
50% true43% positive96% positive
80% true65% positive96% positive
100% true (impossible ceiling)80% positive96% positive

Even the most generous assumption, that every tested hypothesis is true, cannot explain a 96% rate. The literature is not a discovery stream. It is a filter. Either the null results are being systematically buried, or the data is being sculpted to fit the filter. There is no third option.

Sterling (1959) found 97.3% positive in four psychology journals. The same journals, thirty-six years later: 93–99% (Sterling, Rosenbaum, Weinkam 1995). Across twenty disciplines, Fanelli (2010) measured economics at 88.5%, psychology and psychiatry at 91.5%. Between 1990 and 2007, the rate grew another 22% (Fanelli 2012). Ioannidis (2005) gave the Bayesian version: given typical power and pre-study odds, the positive predictive value of a published biomedical finding is below 50%.

The system does not require red-teaming before publication. Everybody is writing the paper where the prediction hit. Science corrects for everything but itself.

The downstream reach is policy. A 2012 paper on signing honesty pledges at the top of a form (Shu, Mazar, Gino, Ariely, Bazerman) was cited in the IRS Behavioral Insights Toolkit and absorbed into the nudge-unit literature before its data was shown to be fabricated. Ego depletion, the claim that willpower is a finite resource, became the spine of Baumeister and Tierney’s bestseller Willpower (2011) before the effect failed to replicate across labs (Hagger et al. 2016). Power posing generated more than 70 million views for Amy Cuddy’s TED talk and built a New York Times bestseller (Presence, 2015). Its first author, Dana Carney, later publicly withdrew in 2016, writing “I do not believe that ‘power pose’ effects are real.” Cuddy did not.

Exhibit 3. Without a work log, fraud runs for years

Yoshitaka Fujii fabricated data across 183 papers over nineteen years of peer review in anesthesiology. John Carlisle, a UK anaesthetist, ran a distributional goodness-of-fit test on Fujii’s published baseline statistics and reported the result in Anaesthesia (2012). The journals had no trail to audit against.

Diederik Stapel: 58 retractions in social psychology. Detection came from three junior researchers in his own department who reported him to the department chair. Stapel was dean, so they had to go around him. Haruko Obokata’s STAP cells: retracted within five months after PubPeer commenters and a blog post analyzed the figures; Nature’s editorial acknowledged that “we and the referees could not have detected the problems that fatally undermined the papers.” Francesca Gino: four papers retracted after Data Colada published a forensic analysis of the data files themselves.

What are the chances the caught fraudsters are the only ones frauding?

Peer review is pre-trial screening. It filters reasoning errors. It cannot catch perjury. A trial without a trail isn’t a trial at all.

Fujii’s fabrications contaminated two decades of anesthesia meta-analyses. Paolo Macchiarini’s synthetic trachea transplants, published in The Lancet and celebrated on television, killed most of their recipients before Swedish journalists forced the scandal open. Wakefield’s 1998 MMR paper coincided with UK vaccination rates falling from ~92% (1995–96) to 79.9% (2003–04), well below the 95% herd-immunity threshold, and seeded outbreaks in US pockets a decade later. The public-health damage followed cultural transmission, not the retraction notice.

Exhibit 4. The citation graph doesn’t update

Serra-Garcia and Gneezy (2021) tracked citations in psychology and economics after replication failure. Papers that failed to replicate were cited more than papers that replicated, and only 12% of post-failure citations mentioned the failure.

Hsiao and Schneider (2021) ran the same question across biomedicine. Of 13,252 citation contexts referencing 7,813 retracted PubMed papers, 5.4% acknowledged the retraction.

The trail isn’t just incomplete. The reader has no reliable way to tell whether the citation they just followed is still standing. Retraction is supposed to be a motion to strike; the citation graph never grants it.

That matters because the reader is often a textbook, a policy brief, a clinical guideline, or a language model. Claims enter downstream corpora and compound. When the original retracts, the downstream copies don’t. The honesty-pledge result kept training students and shaping forms for years after its data was shown to be fabricated, because the citation graph has no reverse gear.

The view from inside

I once watched The Act of Killing, a documentary in which perpetrators of the Indonesian mass killings cheerfully reenact their murders for the camera. From inside the culture that celebrated them, the evil was invisible. They posed, they laughed, they corrected each other’s technique.

The structure transfers even if the moral weight does not. A practice that looks pathological from outside is unremarkable from inside, because the local rituals make sense to the people performing them.

The invisibility runs to the reader. Appeal to authority is a named fallacy. Did you ever believe someone because they studied at Harvard or Stanford? Did you find a paper more credible because it had many citations? If so, you’re a part of the problem. That reflex, unnoticed and routine, is the symptom under diagnosis.

Not every appeal to authority is a fallacy. For a layperson deferring to expert consensus, it is a reasonable division of epistemic labor; you can’t personally check every claim. For one scientist citing another scientist’s credence, when both have the tools to verify, it is a fallacy masquerading as evidence. You have been trained to treat the second as the first.

The same invisibility produced eugenics inside the institution this post is arraigning. R. A. Fisher, patron saint of statistical inference and inventor of the p-value, was an active eugenicist. Peer-reviewed journals published the work through the 1930s, and forced sterilizations continued into the 1970s. The institution now treats eugenics as an alien moral failure rather than as work once published, funded, taught, and credentialed from inside the house. Credentialed consensus, journals, and prestige launder ideology as knowledge. That is the mechanism.

PhD students aren’t villains for wanting to be credentialed. Advisors aren’t villains for teaching them which documents to write and which people to network with. Labs aren’t villains for optimizing publication count. The institution is internally coherent. Whether it is doing science is a question the rituals cannot ask. A discipline that can’t discipline itself is a profession.

Blame the game, not the player.

Who is the jury?

If the jury is other credentialed scientists, science is a priesthood with extra steps. The same people who wrote the papers adjudicate the papers.

The outside jury is already in session. Retraction Watch logged 140 retractions in 2000 and more than 10,000 in 2023. Elisabeth Bik has surfaced image manipulation in more than 4,000 papers. Data Colada’s forensic analyses took down Ariely’s 2012 honesty paper and four of Gino’s; Harvard revoked Gino’s tenure in May 2025, apparently the first such revocation since at least the 1940s. Gino’s $25 million defamation claims against Data Colada were dismissed in September 2024.

In March 2020, Bik publicly criticized Didier Raoult’s hydroxychloroquine paper, flagging non-randomized controls and six omitted treated patients, including one who died. In April 2021, Raoult’s IHU in Marseille filed a criminal complaint against her for harcèlement moral aggravé, tentative de chantage, and tentative d’extorsion. The prosecutor reportedly closed it in March 2024. The paper was retracted in December 2024.

The jury is anyone who can read the evidence record. What it needs is the trail.

It’s a club

Put in economic terms, the scientific community functions as a cartel. Peer review is the admissibility rule. Citation is the currency. Credentialing is the output the members license. When an outside auditor produces an inspectable claim the club cannot contain (Bik, Wilmshurst, Data Colada), the club answers with criminal complaints, libel suits, and defamation claims. Carlin put it more cleanly than any economist: “It’s a big club, and you ain’t in it.”

The journal isn’t a stream of trials; it’s a trophy case for grant applications. What can’t be displayed isn’t submitted, and what is submitted is written for the case.

The labor structure confirms the shape. NIH predoctoral stipends work out to $13.84 an hour at forty hours, below the minimum wage in New York, San Francisco, or Boston. Research hours are rarely forty. Postdocs at the NIH scale approach or cross the line at 60–70 hour weeks. 14% of doctorate recipients self-finance out of personal resources. Only 7–8% of 2024 research-doctorate recipients had an immediate tenure-track job lined up. The lotto rarely pays out.

The price of the ticket also rises. Academic postdocs grew from 12,500 in 1979 to 70,000 in 2024; multi-year postdoc chains are standard, and assistant-professor candidates arrive at hiring with materially more publications than their predecessors. The cartel’s temporal shape is a pyramid scheme. Early entrants cashed in. Later cohorts run on inflating credentials against diminishing real positions, and each new generation subsidizes the last with cheaper labor and heavier dues.

The cartel is not a failure of the institution. It is the institution operating as designed. Good work still gets done inside it, by people paying personal cost to do it right.

”But Gino and Ariely were outliers”

Three of the four exhibits above describe honest work. Exhibit 1, where 49 percentage points of positive cardiovascular findings evaporated under pre-registration, involves no fraud. Exhibit 2, the 96% positive-result rate persisting since 1959, requires no fabrication. P-hacking, HARKing, and selective outcome reporting are all within the rules. Exhibit 4, citations that don’t update when papers retract, is infrastructure, not malice.

Fraud is the one failure mode the institution acknowledges. The other three are what happens when the protocol is optional.

The 96% rate predates Ariely’s first paper by half a century. Sterling documented 97% positive in 1959. Gino was not yet alive.

”Science is self-correcting”

If self-correction means publishing a retraction, yes. If it means the update propagating through the literature, no.

Hsiao and Schneider found that 5.4% of citations to retracted biomedical papers acknowledge the retraction. Serra-Garcia and Gneezy found that papers that failed to replicate are cited more than papers that did. Retractions are filed, not enforced.

A self-correcting system requires that the correction propagate faster than the original claim. The citation graph runs the other way.

The record also fails to carry the corrections the field already knows about. Replication failure is not a retraction criterion, so retractions alone won’t match replication rates. But the literature should carry visible caution markers (failed-replication notices, effect-size revisions, boundary-condition warnings) on papers whose claims the field has stopped believing. It doesn’t. A paper that stopped replicating a decade ago is still cited naked, indistinguishable from one that held up. The gap between what the field knows and what the record shows is two orders of magnitude.

”Failed replication doesn’t mean the original was false”

A result that does not transport has not earned the general claim built on top of it. Pearl’s point: transportability requires explicit assumptions about causal structure and context, assumptions usually unstated and untested. If your finding holds only for college sophomores on a Tuesday in 1998, you didn’t discover a law of human nature. You recorded an anecdote. Publishing it as a law, and citing it as if it held, is the error.

”Pre-registration isn’t appropriate for exploratory science”

It is, and especially then. A declared goal shapes the exploration. Science without a goal is philosophy. Registering the question is the courage to be publicly wrong, and it coordinates other scientists who can shake maximal surprise out of the data instead of settling for a 95%-confidence prior. Exploratory registration isn’t a hypothesis lock; it’s a record of what was being looked for, so the reader can distinguish what the researcher expected from what the data said.

”Truthseeking happens despite the institution”

Yes. That is the argument. Real discovery is already routing around the credentialing mechanism, inside universities as much as outside them.

Science is happening. The fastest correction mechanisms are often outside the journal process.

”The outside jury can become a mob”

Yes, outside scrutiny can become performative or punitive. Online auditors can be wrong; reputational damage can run ahead of adjudication; selective outrage is real. That is why the standard has to be inspectable claims, not vibes or pile-ons. What the outside jury offers, at its best, is exactly that: the work is checkable, and the checker is checkable in return. The inside, by contrast, has been publishing mob verdicts, credentialed and peer-reviewed and catastrophically wrong, for decades. Eugenics, lobotomies, recovered memory, cold fusion, ego depletion.

Peter Wilmshurst, co-principal investigator of the MIST cardiac-device trial, publicly challenged the trial’s handling at a 2007 conference and refused to sign off on the Circulation paper. NMT Medical, the sponsor, sued him for libel in English court. In December 2010 the High Court ordered NMT to post £200,000 in security for costs; NMT went into liquidation in April 2011 and the case died with it.

A noisy but auditable process is already more accountability than the inside offers.

”Without peer review, how am I supposed to trust anybody?”

Credentialed consensus is how geocentrism survived. For centuries, Europe’s credentialed natural philosophers inherited a geocentric cosmology and defended it against heresy. The consensus broke not because of better review but because Kepler had data and Galileo refused to defer.

The standardized, journal-centered, anonymous external referee process now treated as the source of scientific trust is largely a twentieth-century consolidation. Darwin, Einstein, Newton, Maxwell, Mendel, Watson and Crick: none of their canonical work passed through today’s referee regime. Was science untrustworthy until the mid-twentieth century?

The objection itself is an appeal to authority. “How do I trust anybody” asks for a credential to defer to. That is the pathology, not the remedy. What you need is not another authority; it is the ability to check.

Peer review was never the source of trust. It is pre-trial screening. What produced trust in science was prediction success, replication, and engineering validation. Reality pushing back on bad claims, regardless of what the journals said. You trust physics because rockets don’t explode, not because Physical Review approved the equations.

What to trust instead:

Peer review catches reasoning errors. Useful, not dispositive, insufficient. The institution lost the thread when it started treating it as a certificate of truth.

”Reform is already happening”

Some reform, yes. The incentives are not. Registered Reports exist but are not the default career currency. Pre-registration is available but optional. Data-sharing mandates are policy but rarely audited. Follow the money: funding, promotions, and tenure still flow to publication count, not to trail quality. None of this evidence is new: Sterling published in 1959, Ioannidis in 2005, the Open Science Collaboration’s 97% → 36% replication rate landed in 2015, Registered Reports have cut positive rates in half wherever tried. The institution has absorbed each wave without updating. Reform without incentive change is sanitation theater.

”It’s the best we can do”

The argument: reviewers have finite bandwidth, editors have deadlines, and scientists cannot be expected to document every false start while producing new work.

Every other industry that does serious work refutes this. Software engineers run version control, code reviews, and postmortems. Aviation maintains incident logs and black boxes. Pharmaceutical labs keep rigorous notebooks because patent law requires it. Finance runs on audit trails. Medicine runs on surgical checklists and M&M conferences. None publish externally. All maintain internal trails that survive the individual’s departure.

Publishing the protocol does not mean announcing results to the world. It means committing to a trail the next person in your role can audit. The scope of that trail is whatever IP, regulation, or competitive position allows: private, classified, internal. The discipline is not optional. The academic institution claims the name of science while leaving its trails less audited than any industry where money or lives are at stake.

The jury’s charge

The exhibits close. The jury’s job begins.

Ask whether you can reproduce what you cite. Reject credence on credence alone; demand the evidence, the methodology, the thought process that led to the conclusion. Follow the outside jury: Retraction Watch, Data Colada, Bik, PubPeer. Fund them where you can. Learn the protocol’s four-century history. Each step was forced by a prior failure.

If you write science: pre-register, publish the nulls, keep the log, release the data, write the replication. Run every draft through an LLM adversary first: every citation, every claim, every inference. If the LLM can find the hole, the replicator will.

If you fund or hire: trail quality as the metric. Mandate data sharing. Require Registered Reports for confirmatory work. Pay for replication.

The reforms exist. The incentives don’t.

My own exhibit

I have done this. In a preregistered experiment I ran, the final posterior came in at 0.949 — 0.001 below the 0.95 threshold I had committed to beforehand. By my own rules, the result does not confirm. I said so. It was harder than any positive result I have ever published. This is the flex. Your turn.


Written via the double loop.