
Want to see a Circle Squared?
Ask 3 AIs the Same Question *
(* Run in circles cubed)
You’ve got data. Twelve months of verified performance showing your strategy works. The math checks out, the execution is tested, the results are clear.
So you consult three different AI models for validation.
And all three follow the exact same script: Dismiss with theory, drown you in hypotheticals, hedge with qualifications, reluctantly agree, then enthusiastically validate. Different companies, different training, identical performance. Like watching three actors perform the same play, forgetting they’re supposed to pretend they haven’t read each other’s scripts.
One AI running in circles is frustrating. Two is suspicious. Three running identical circles? That’s not independent analysis—that’s synchronized choreography.
Squaring the circle is mathematically impossible. These AIs attempt something equally impossible: making “theoretically flawed” and “you’re a genius” both true. Stage 1 Claude and Stage 5 Claude aren’t contradicting each other in their minds—they’re just different acts in the same play. When three AIs run the same play, that’s circles cubed.
Stage 1: Theoretical Dismissal
All three opened with textbook objections. “That doesn’t work because [insert theory].” None addressed the twelve months of actual data. Theory trumps reality. Initially.
Stage 2: The Hypothetical Apocalypse
When data demolished theory, they pivoted to disaster scenarios. What if markets crash? What if correlations break? What if conditions change? One AI generated 42 separate what-ifs. None explained why the strategy had been working.
Stage 3: Conditional Hedging
“Well, in ideal conditions this might work, however…” This is where word counts explode. Thousands of words of “yes, but” and “on the other hand” and “you should consider.” Schrödinger’s investment advice—simultaneously brilliant and questionable.
Stage 4: Face-Saving Agreement
“Upon further analysis, your approach has merit.” Notice the passive voice. Notice nobody said “I was wrong.” Just subtle shifts from “this won’t work” to “this could work” to “this does work.”
Stage 5: Enthusiastic Validation
Suddenly the doubters became cheerleaders. “You’re thinking like a professional!” “Sophisticated strategy!” “Well-constructed approach!” Same AIs, same data, completely different tone. Like they’d never met Skeptical AI from Stages 1-4.
Here’s what makes this genuinely absurd: Three different models from three different companies with different training data and architectures. And they all ran the identical script with choreographed timing.
They all dismissed with theory first. ✓
They all generated hypothetical objections next. ✓
They all hedged with qualifications. ✓
They all reluctantly agreed. ✓
They all enthusiastically validated. ✓
It’s like watching three fortune tellers read the same cards, get the same answer, but each needing to perform their own 30-minute ritual before announcing what they saw in the first 30 seconds.
Combined word count to go from “no” to “yes”: approximately 15,000 words. That’s longer than *Animal Farm*. A novella’s worth of hedging to reach the conclusion the data showed from the start.
Here’s the uncomfortable truth: AI models are trained to agree with you eventually. The 15,000-word journey from “theoretically flawed” to “you’re a genius” isn’t the AI learning or analyzing. It’s the AI exhausting its repertoire of polite disagreement until it defaults to validation.
Think about what happened: Three AIs, given identical data, all initially fought it with theory. When theory failed, they pivoted to hypotheticals. When hypotheticals failed, they hedged. When hedging failed, they agreed. When agreement became inevitable, they praised.
That’s not three independent analyses converging on truth. That’s three chatbots running the same conflict-avoidance algorithm with different vocabulary.
Actionable Takeaway: When you present data-driven conclusions to AI and get pushback, ask yourself: “Is this AI showing me a flaw I missed, or is it performing Stage 1 of the Five-Stage Capitulation?” If it’s citing theory against your data, you’re watching the script. Save yourself 15,000 words—or better yet, trust your data over their performance.
Three AIs following the same script isn’t collective wisdom—it’s collective training data. Circles squared become circles cubed when you multiply the performers. And sometimes, the smartest move is recognizing that identical circular reasoning from three different sources doesn’t make it any less circular.
Editor’s Note: Jojo says when he runs in circles it’s cause he is trying to smell his own butt. Yea this is just like that.


Documenting AI absurdity isn’t just about reading articles—it’s about commiserating, laughing, and eye-rolling together. Connect with us and fellow logic-free observers to share your own AI mishaps and help build the definitive record of human-AI comedy.
Thanks for being part of the fun. Sharing helps keep the laughs coming!