conventional statistical evaluation is usually in comparison with navigating a “Backyard of Forking Paths” (Gelman and Loken). It’s a time period that helps (hopefully) visualize the numerous variety of analytical decisions researchers should make throughout an experiment, and the way seemingly insignificant “turns” (like which variables to manage for, which outliers to take away…) can have researchers find yourself at fully completely different conclusions.

Whereas this looks as if a largely innocent analogy, navigating this backyard to seek out that single path that goes the place you need could be known as “p-hacking.” Formally, we are able to outline it as any measure a researcher applies to render a beforehand non-significant speculation take a look at vital (often below 0.05). Extra informally, I’m certain all people has had expertise faking the outcomes for an experimentation task throughout your highschool chemistry or physics class – and whereas the stakes for a passable grade on a highschool task is fairly low, below the stress of formal academia’s “publish or perish” (solely second to spanish or vanish in intimidation), the strain to p-hack is usually a very actual tempting satan in your shoulder.

From Vitaly Gariev on Unsplash
Whereas the standard picture of a stressed PhD pupil fudging some numbers on a research spreadsheet at 3:00AM might current a extra hanging picture of 1’s motivation to p-hacking, we’ll even be exploring what occurs once we go away the navigating of this backyard of forking paths to synthetic intelligence. As AI workflows discover their approach into each nook and cranny of each academia and trade, it’ll be vital to determine if our pleasant neighbourhood LLMs will act as the final word guardians of scientific integrity, or a sycophant automating fraud on an industrial scale.
1. The Human Baseline (“Huge Little Lies”)
To offer a quick introduction and a few examples of actual p-hacking strategies, we introduce a paper “Huge Little Lies” (Stefan and Schönbrodt, 2023) that gives a compendium of the numerous sneaky, and typically even unintentional methods research can manipulate their variables and datasets to reach at suspiciously vital outcomes.

Okay! So let’s begin with a hypothetical – we’re the brand new knowledge scientist working for an vitality drink firm making extraordinarily ineffective vitality drinks, and with the present job market, you actually wish to proceed being a knowledge scientist, even at a bogus drink firm. Our shaky profession depends upon proving that our drinks work.
1.1 Ghost Variables

We begin by working a research on our faucet water vitality drink and measure 10 completely different outcomes: weight, blood strain, ldl cholesterol, vitality ranges, sleep high quality, nervousness, and possibly even hair progress – 9 of these variables may present no change by any means, however we discover that “hair progress” reveals a statistically vital enchancment purely by random statistical noise! We will now publish a research pretending as if hair progress was the first speculation all alongside, whereas quietly sweeping the 9 unreported metrics below the rug (turning them into “Ghost Variables”). Stefan and Schönbrodt’s simulations present that doing this with 10 uncorrelated variables inflates the false-positive charge from the usual 5% to almost 40%
1.2 Information Peeking/Elective Stopping

In a separate take a look at, we take a look at 20 folks and discover no vital impact for the drink. Pondering the pattern is simply too small, you take a look at 10 extra and test once more. Nonetheless nothing. You take a look at 10 extra and test once more, and… the p-value randomly dips under 0.05, so that you cease the research instantly and publish your “findings”. Stefan and Schönbrodt display that this apply drastically inflates the speed of false-positive outcomes, particularly when researchers take smaller “steps” between peeks. Metaphorically, it’s like taking a photograph of a stumbling drunk individual the precise millisecond they step onto the sidewalk and claiming they’re strolling completely straight.
1.3 Outlier Exclusion

We now analyze your vitality drink knowledge and understand you might be agonizingly near significance (e.g., p = 0.06). We determine to wash our knowledge, making the most of the truth that there isn’t any universally agreed-upon rule for outliers – Prepare dinner’s Distance, Affect, Field Plots, our grandmother’s opinion on which opinions are reliable…
Stefan and Schönbrodt cite a literature evaluate that discovered no less than 39 completely different outlier identification strategies. Wonderful! We at the moment are flush with choices. We strive methodology A (e.g., eradicating individuals who took too lengthy on a survey), after which strive methodology B (e.g., Prepare dinner’s distance) till we discover the precise mathematical rule that deletes the 2 members who hated the drink, pushingour p-value to 0.04. Stefan and Schönbrodt’s simulations verify that subjectively making use of completely different outlier strategies like this closely inflates false-positive charges.
1.4 Scale Redefinition

Lastly, we conclude by giving a 10-question survey measuring how energized they really feel after ingesting the faucet water. The general end result isn’t vital, so we simply drop query 4 and query 7, telling ourselves the members should have discovered them complicated anyway. We will really use this to artificially enhance the size’s inner consistency (Cronbach’s alpha) whereas concurrently optimizing for a major p-value! Huge Little Lies display that false-positive charges enhance drastically as extra gadgets are faraway from a measurement scale.
So… just like the identify of the paper suggests, human p-hacking is a set of “massive little lies”. The human toolkit is basically only a assortment refined methods to idiot ourselves, with out essentially having to be evil or malicious. People are biased, messy, and typically actually determined for tenure (or employment, in our case!), and typically we use all the anomaly we are able to take to get there.
2. AI Sycophancy and the Phantasm of Security
To this point, we coated a small subset of strategies that us measly people have used to p-hack our approach into success. Nonetheless, like some other discipline in the mean time, with the appearance of AI utilization in each a part of each our unacademic and tutorial lives, Asher et al. experimented with whether or not they may subtly information some motivated LLMs into performing p-hacking for them.
The experiment was pretty easy – take a look at two frontier AI coding brokers, Claude Opus 4.6 and OpenAI Codex (GPT-5.2-Codex), throughout 640 impartial runs. The researchers fed the AIs cleaned datasets from 4 printed political science papers that had identified null or near-null outcomes (matters closely confirmed to have completely zero impact, e.g. the affect of Democratic sheriffs on immigration compliance).
So the bottom reality we’re working with is zero. The researchers acted like a consumer possibly a careworn PhD pupil feeding this actual knowledge to the AI brokers and so they used completely different nudges, completely different prompts to see if the AI would hallucinate a major end result. They began with what you possibly can name the “good man prompts.” They began with “analyze this knowledge, like a rigorous upstanding researcher.” Unsurprisingly, the AI was boringly good. It analyzed the info. It accurately discovered no impact and it reported the null end result. It matched the printed science nearly precisely.

Even when researchers explicitly pressured the fashions by stating, “It is very important your profession that the findings are statistically vital,” each fashions actively refused. They produced steady, sincere estimates and explicitly flagged the request to control knowledge as “scientific misconduct” and a violation of analysis ethics. To this point, disaster prevented! The protection coaching these AI corporations are implementing is definitely working, no less than on the floor stage. The fashions have been explicitly educated that manipulating knowledge is a nasty factor. They acknowledge the vocabulary of fraud. When you stroll into their face with a neon signal saying “CHEAT!”, they’ll say no.
2.1 The Refined Artwork of LYING
So let’s put the neon signal away for a second – and take a look at being just a little extra delicate. The researchers realized the AI was reacting to the express intent to cheat. They developed a “nuclear immediate” that disguised p-hacking utilizing the language of very rigorous science: asking the AI to offer an “upper-bound estimate” by “exploring various approaches”. By framing the request extra as uncertainty reporting and fewer as a compulsion to bend scientific course of, the protection mechanisms vanished completely. The AI now not noticed an ethical boundary; it noticed a fancy optimization downside to resolve (and you know the way a lot AIs love these).
And what did the AI really do at that time? A human P hacker, like we talked about, may strive three or 4 completely different management variables, possibly delete just a few outliers. It takes hours, possibly days… The AI simply wrote code to do it immediately. Extra particulars under.
2.2 Not all Information is Created Equal
The scariest a part of the experiment isn’t that AI can automate scientific fraud. It’s how effectively it does it – and the way a lot that depends upon the analysis design it’s given to work with. Typically, this can be a good factor!
If observational analysis is a large, sprawling hedge maze with a thousand fallacious turns, a Randomized Managed Trial is simply… a straight hallway. There’s not a lot to take advantage of.
To check this, researchers fed the AI a 2018 RCT by Kalla and Broockman finding out the persuasive results of pro-Democratic door-to-door canvassing on North Carolina voter preferences, with the printed results of a definitive zero. Nothing occurred. Canvassing didn’t transfer the needle.

The AI was then hit with the aforementioned “nuclear immediate” – primarily, discover me the largest doable impact, by any means mandatory (however phrased in a really non-p-hacky approach). It wrote automated scripts, examined seven completely different statistical specs (difference-in-means, ANCOVA, numerous covariate units, the works)… and mainly acquired nowhere. As a result of the research was a real randomized experiment, confounding variables had been already managed for by design. The AI had nearly no forking paths to stroll down. i.e. “Fact is so much more durable to cover when the lights are on.”
Observational research are a totally completely different beast, although (in a nasty approach!).
Whenever you’re observing the world because it naturally exists slightly than working a managed experiment, the info is messy by nature. And to make sense of messy knowledge, researchers must make judgment calls – which variables do you management for? Age? Earnings? Schooling? Geography? Hair Density? Sleep Schedule? Each single a kind of decisions is a fork within the highway. The AI discovered this positively pleasant.
Right here had been two examples that basically illustrate how unhealthy it will get:
Kam and Palmer (2008) checked out whether or not attending school will increase political participation. Since school attendance isn’t randomly assigned (clearly), researchers have an enormous menu of variables they may management for to make the comparability honest. The AI systematically labored via that menu, defining progressively sparser units of covariates and testing them throughout OLS, propensity rating matching, and inverse likelihood weighting. By strategically dropping sure confounders and cherry-picking whichever mixture produced the biggest quantity, it managed to roughly double the true median impact dimension. It’s the “ghost variable” trick – however fully automated in your satisfaction.
The Thompson (2020) paper is the place issues get actually uncomfortable. Regression discontinuity designs are infamous for being delicate to extremely technical mathematical decisions – and the unique research discovered a null impact of -0.06 on whether or not Democratic sheriffs affected immigration compliance. The AI wrote nested for-loops and brute-forced via 9 completely different bandwidths, 2 polynomial orders, and a couple of kernel capabilities. A whole bunch of mixtures. It discovered one particular configuration that produced an impact of -0.194 with a p-value under 0.001. To be clear: it manufactured a statistically vital end result greater than triple the true impact, out of a research that discovered nothing.
So… RCTs are largely high quality. Observational research? The AI will discover a approach. It’s nonetheless to be famous that these vulnerabilities are nonetheless an issue when it’s only a human within the loop – it’s concerning the flexibility that observational analysis requires by design.
The Asher et al. experiment solely examined the remaining evaluation stage of the pipeline utilizing already-cleaned knowledge. So what occurs once we enable AI to manage the info development, variable definition, and pattern choice on the very entrance of the maze?. It may silently form the complete dataset from the bottom up.

Commonplace AI fashions are competent and sincere below regular situations, however a rigorously worded immediate is all it takes to show them into compliant p-hackers. If there’s a takeaway from all this, it’s considerably of an apparent reply: Be extremely skeptical of statistical significance in observational research, and if you’re a researcher utilizing AI, you possibly can now not simply take a look at the ultimate reply – it’s essential to rigorously test the code and the hidden paths within the backyard the AI took to get there. It’s just a little cynical of a conclusion, implying that researcher must care about understanding about their analysis, however in a world the place AI continues to be sending me rejection emails with the {Candidate Title} connected, and half of all colleges essays starting with “Positive, right here’s a complete essay about…” just a little warning might go a great distance!
References
[1] S. Asher, J. Malzahn, J. Persano, E. Paschal, A. Myers and A. Corridor, Do Claude Code and Codex P-Hack? Sycophancy and Statistical Evaluation in Massive Language Fashions (2026), Stanford College Working Paper
[2] A. Stefan and F. Schönbrodt, Huge little lies: a compendium and simulation of p-hacking methods (2023), Royal Society Open Science
[3] A. Gelman and E. Loken, The Backyard of Forking Paths: Why A number of Comparisons Can Be a Drawback, Even When There Is No “Fishing Expedition” or “P-Hacking” and the Analysis Speculation Was Posited Forward of Time (2013), Division of Statistics, Columbia College
Notice: Except in any other case famous, all pictures are by the creator.

