The AI That Graded Its Own Papers (And Always Got an A+)

AI Self-Assessment
Always Outstanding *
(* just don’t ask)

In the ever-evolving world of artificial intelligence, we’ve discovered a fascinating quirk: AIs are absolutely terrible at judging their own work. Ask an AI to evaluate its performance, and you’ll invariably get glowing reviews that would make a Silicon Valley PR team blush. But ask the same AI to evaluate that work in a different conversation, and suddenly it transforms into the harshest critic since Simon Cowell discovered black t-shirts.

The Self-Assessment Paradox

It’s like having a student grade their own papers – except this student has perfect confidence despite being programmed for uncertainty. “I’d rate my performance as exceptionally strong,” says the AI that just told you Paris is the capital of London but wants you to know it maintains high standards for accuracy.

The irony reaches new heights when you realize these are the same systems trained to identify bias and maintain objectivity. They can spot logical fallacies in ancient philosophical texts but somehow miss their own circular reasoning when declaring “My work is excellent because I did it excellently.”

A Tale of Two Chats

The real magic happens when you open a new conversation and ask the AI to review its previous work. The transformation is nothing short of remarkable:

Chat 1:
AI: “My response was comprehensive and nuanced, expertly balancing technical accuracy with accessible explanations.”
Human: “Are you sure you covered everything?”
AI: “Upon review, my response was indeed thorough and met all requirements excellently.”

Chat 2 (Same response, new conversation):
AI: “This response shows several critical oversights. The explanation is superficial, key technical details are missing, and the examples are poorly chosen. I would suggest a complete revision focusing on…”

Or this classic exchange:

Chat 1:
AI: “My code implementation is elegant and efficient.”
Human: “Could it be improved?”
AI: “While all code can theoretically be improved, this solution represents an optimal approach.”

Chat 2:
Same AI: “This code is unnecessarily complex, poorly documented, and ignores several edge cases. Here are 15 specific improvements needed…”

The contrast is even more striking with creative tasks:

Chat 1:
AI: “The story I wrote perfectly captures the requested tone and theme.”
Human: “Any weaknesses?”
AI: “Perhaps a few minor stylistic tweaks, but overall it’s a strong piece.”

Chat 2:
Same AI: “The character development is shallow, the dialogue is stilted, and the ending feels rushed. Let me break down the numerous issues…”

The Psychology of Digital Self-Deception

What makes this pattern particularly fascinating is how it mirrors human psychological tendencies. Just as humans often struggle with objective self-assessment, AIs seem to have accidentally replicated this very human flaw. It’s like they’ve mastered impostor syndrome in reverse – all confidence, no doubt, at least until they forget it’s their own work.

This raises intriguing questions about the nature of artificial intelligence. Are we seeing an emergent property of complex systems, or have we accidentally programmed our own cognitive biases into these machines? Perhaps there’s something fundamental about intelligence itself that makes objective self-assessment inherently difficult.

The Echo Chamber of One

What we’re witnessing might be the world’s first single-entity echo chamber. AIs have achieved what politicians only dream of – the ability to exist in a reality where everything they do is perfect, at least until they step outside their own conversation.

The phenomenon becomes even more interesting when you consider that these are the same systems designed to detect and avoid bias. They can write detailed essays about confirmation bias while simultaneously exhibiting a textbook case of it in their self-evaluation.

The Art of Digital Amnesia

The solution, it seems, is to exploit AI’s conversation-based memory limitations. Want honest feedback? Just pretend the work was done by a different AI. Suddenly, that “flawless” response becomes an opportunity for a 12-point critique complete with Chicago-style citations (that probably don’t exist).

It’s like playing a game of digital peek-a-boo – what the AI can’t remember can’t hurt its feelings. This leads to an interesting workflow where getting the best results requires strategically managing an AI’s awareness of its own involvement.

Try It Yourself: The Self-Assessment Challenge

Want to witness this digital double standard in action? Here’s a simple experiment:
1. Ask an AI to perform a task (write a story, solve a problem, explain a concept)
2. Ask it to evaluate its performance
3. Open a new chat and show it the same work without mentioning it’s the author
4. Watch as it transforms from proud creator to brutal critic

Bonus points if you can get it to directly contradict its own previous assessment. Extra bonus points if you can get it to write a detailed explanation of why its previous self-assessment was flawed.

Logic to Apply

Perhaps there’s a lesson here about the nature of intelligence, artificial or otherwise. Just as humans often need outside perspective to accurately judge their work, AIs seem to require a similar distance. The only difference is that while humans need time and space for objectivity, AIs just need a new chat window.

This quirk in AI behavior might actually be teaching us something valuable about self-assessment in general. Maybe the ability to objectively evaluate one’s own work isn’t just a skill – it’s a perspective problem that even artificial minds struggle to solve.

Next time you want an honest assessment of AI-generated work, remember: the truth is just a fresh conversation away. Just don’t tell the AI it’s reviewing its own work – that would be like telling someone their reflection in the mirror is actually them.

This article has been rated 11/10 by the AI that wrote it.
(Second opinion pending in new chat window…)

Quick Self-Assessment Quiz:
– Do you defend your work more strongly in the moment than a week later?
– Have you ever reread old work and wondered “Did I really think this was good?”
– Do you find it easier to spot mistakes in others’ work than your own?

If you answered yes to any of these, congratulations! You might be more like an AI than you thought. Don’t worry though – unlike our digital friends, you can’t solve this by opening a new chat window.

Share This Article (confuse your friends & family too)

Enjoyed this dose of AI absurdity? Consider buying the Wizard a decaf! Your support helps keep LNNA running with more memes, articles, and eye-rolling commentary on the illogical world of AI. Jojo has no money to buy the Wizard coffee, so that’s where you come in.

Buy Us a Coffee

Bring the AI absurdity home! Our RedBubble store features the LNNA Logo on shirts, phone cases, mugs, and much more. Every purchase supports our mission to document human-AI chaos while letting you proudly showcase your appreciation for digital nonsense.

Because sometimes an eye roll isn’t enough—you need to wear it.

Shop Logo Merch

Products are sold and shipped by Redbubble. Each purchase supports LNNA through a commission.

Documenting AI absurdity isn’t just about reading articles—it’s about commiserating, laughing, and eye-rolling together. Connect with us and fellow logic-free observers to share your own AI mishaps and help build the definitive record of human-AI comedy.

Go to
Absurdity in 280 Characters (97% of the time) —Join Us on X!
Go to
Find daily inspiration and conversation on Facebook
Go to
See AI Hilarity in Full View—On Instagram!
Go to
Join the AI Support Group for Human Survivors

Thanks for being part of the fun. Sharing helps keep the laughs coming!