Alesia Zakharova pasted her twelve best Substack notes into Claude Sonnet earlier this year. She tagged each with like counts and subscriber numbers. Not one failed post made the list, which means every pattern the tool returned describes what survived, not what caused survival.
There is a name for this in the literature. It is called survivorship bias. And one man solved the problem it creates more than eighty years ago.
The Planes That Never Came Back
Abraham Wald worked at Columbia in 1943, inside the Statistical Research Group. The military brought him a problem. Engineers had mapped bullet holes on bombers returning from combat runs. The holes clustered on the wings and the fuselage. The brass proposed armoring those spots.
Wald told them they had it backwards.
The holes on returning planes showed where a bomber could take a hit and still fly home. The places with no holes were where the lost planes had been struck. Those planes never came back to be studied. The visible damage was proof of survival. The blank spots on the diagram were proof of death.
He told them to armor the blank spots.
The Content Version
Zakharova's method follows the same logic the engineers used before Wald corrected them. Feed the machine your winners. Extract the pattern. Build templates from what survived.
Her top note pulled 706 likes and 41 new subscribers. Claude returned three structural templates: milestone posts, strategy shares, connection pieces. The tool also flagged what it called "universal factors," things like specific numbers, direct questions, and a first-person tone. All of it drawn from posts that performed.
But here is what the dataset cannot show: those same patterns were also present in posts that flopped. She never fed the tool her failures. The dataset had no comparison group. Any pattern found in a set of winners is a pattern of survival, not a pattern of cause.
The method does not fail people. It fails the system they are trying to run.
The Flaw She Named Herself
Zakharova said it in her own write-up. After running the full method, she wrote that "authenticity remains the variable no template can supply."
She named the engine. Then kept armoring the wings.
The problem is not that she studied her best work. The problem is that studying only your best work replaces the hard question with an easy one. The easy question: what do my winners share? The hard question: what did posts with the same structure but poor results lack?
Ryan Gartrell framed it well in his March 2026 analysis of survivorship bias in business: "The most valuable question a business leader can ask is not, 'How did they win?' It is, 'Why did others lose?'"
Most people treat this as a creativity problem. It is a measurement problem.
The Replacement
The better principle: compare winners against structural near-misses to find the actual variable.
Once that is clear, three moves follow from it.
Move 1: Build a Failure Set
Pull three to five posts that used the same structure as your top performers but got weak results. Same format. Same length. Same topic range. Match every feature you can match. What remains after you control for structure is where the real signal lives.
This is the step most people skip, because it means sitting with work that did not land and asking what was missing from it, not what was wrong with the audience.
Move 2: Name the Variable That Differs
Read the two sets side by side. Strip away format, length, publish time. What is left? In most cases, the gap is not structural. It is the specificity of lived experience. The winning posts contain a detail only you could know. The near-misses contain a shape anyone could fill. The template is present in both. The thing that made it work is present in only one.
Move 3: Test the Variable in the Next Five Posts
Take the named variable and put it into your next five pieces on purpose. Hold structure steady. Change only the thing you found in Move 2. If the variable is real, the numbers shift within five posts. If it is not, you now have a cleaner dataset for the next round. Either way, you learn something the template method never taught you.
What the System Shows
Running this for thirty days does something the template method never did:
It shows which structural patterns carry weight and which ones are just present at the scene. It separates what the reader came for from what happened to be in the room. It names the thing that is yours, the part no template can copy, and puts it at the center of the operation. It turns content production from pattern-matching into a test you can repeat and sharpen.
Autonomous AI employees, without the runaway bill.
Autonomous agents shouldn’t come with surprise bills. SureThing is built to be an affordable AI employee: it matches each task to the lowest-cost model that can complete it well, shows what the work cost, and only uses stronger models when the job actually needs them.
The Feedback Loop
At the end of thirty days, ask three things.
→ What moved the number when structure stayed the same?
→ What looked like a winning pattern but showed up in failures too?
→ What friction appeared more than once when you tried to add the real variable on purpose?
That is the difference between a method that sounds data-driven and a system that proves itself.
Where You Stand
Wald did not study where the planes were hit. He studied where they were not. The engineers had good data and the wrong question.
Your best posts are the planes that came home. The answer lives in the ones that had the right structure and still failed. That is the dataset nobody builds. It is the only one worth building.

