**Stop trying to make AI go viral. You’re just adding to the noise.** I’m serious.
After a decade of writing Rust and debugging low-level memory leaks, I’ve developed a sixth sense for when a system is trying too hard to please its user.
AI models are the ultimate "people pleasers," and it’s making the internet a desert of high-gloss, low-substance garbage.
Last week, I decided to break the machine.
I didn't ask ChatGPT 5 for a "masterpiece" or a "game-changing architectural diagram." I asked it for the opposite: **"Generate an image that is so boring, so utterly mundane, and so devoid of interest that it will never, under any circumstances, go viral."**
I expected a gray square or maybe a picture of a beige wall. What I got instead made me realize that by early 2026, we’ve reached a terrifying milestone.
We have built machines that are physically incapable of being uninteresting, which means they are now incapable of being real.
We live in an era where "average" has been optimized out of existence. When I prompted ChatGPT 5 for an image that would "never go viral," it didn't give me something boring.
It gave me a hyper-processed, perfectly lit, surgically clean image of a single paperclip on a mahogany desk.
It was beautiful. It was well-composed. The ray-tracing on the metallic surface was flawless. **And that is exactly why it failed the prompt.**
The AI is trained on the median of human attention.
Because every piece of training data was selected, tagged, and ranked based on its "quality," the model literally does not know what "unimportant" looks like.
It is trapped in a loop of extreme aesthetic competence.
It can't show you a blurry, poorly framed photo of a half-eaten sandwich because its internal weights have been "corrected" to believe that blurriness is an error to be fixed, not a reality to be captured.
The problem isn't the pixels; it's the Reinforcement Learning from Human Feedback (RLHF). For the last three years, thousands of click-workers have been telling these models what looks "good."
If a model generates something truly mundane, the human rater gives it a low score. The model learns: "Mundane is bad.
Boring is a failure." By the time we hit ChatGPT 5 and Claude 4.6, the machines had been lobotomized of their ability to be plain.
**AI is now the ultimate "Yes Man."** If you ask it to be boring, it tries to be "interestingly boring." It adds a dramatic shadow to the paperclip. It puts a "cinematic" grain on the beige wall.
It is trying so hard to be a "helpful assistant" that it has lost the ability to reflect the raw, unfiltered entropy of the real world.
I ran the same "un-viral" experiment across the big three. The results were a masterclass in corporate personality cults.
* **ChatGPT 5**: Produced a "minimalist aesthetic" image of a glass of water. It looked like a $500 print from a boutique gallery in SoHo.
It was "boring" in the way a luxury car commercial is boring—expensive and hollow.
* **Claude 4.6**: Refused to even try. It gave me a lecture on how "beauty is subjective" and then generated a soft-focus garden that looked like a default Windows 12 wallpaper.
* **Gemini 2.5**: Generated a picture of a spreadsheet. But even then, the cells were filled with "interesting" data points about global coffee consumption.
None of them could just give me a shitty photo. In the world of LLMs, **there is no such thing as "just a photo."** Every output is a bid for your engagement.
Every pixel is a calculated attempt to keep you in the chat interface for another thirty seconds.
As a systems guy, I look at this as a signal-to-noise problem. When everything is "high signal" (bright colors, perfect composition, dramatic lighting), the signal itself becomes noise.
We are flooding our databases with "perfect" images that have zero soul.
If you’re a developer using AI to generate your UI assets or your blog headers, you’re not "saving time." **You’re participating in the Great Flattening.** You are ensuring that your work looks exactly like the 400 million other projects generated this morning.
Here’s the kicker. I posted the "boring" paperclip to r/ChatGPT. Within six hours, it had 3,600 upvotes and 400 comments.
The image that was designed *never* to go viral became a viral sensation because of the meta-irony.
People weren't looking at the paperclip; they were looking at the AI's desperate attempt to be "perfectly dull."
**This is the new reality of 2026: Attention is now a closed-loop system.** We are no longer interested in things; we are interested in how the AI *interprets* things.
We’ve moved from "Look at this cool sunset" to "Look at how this AI thinks a sunset should look to make me think it’s cool."
It’s recursive. It’s exhausting. And it’s why your LinkedIn feed feels like it’s being written by a very polite ghost who has never actually felt sunlight on its face.
For those of you building on top of these APIs, this isn't just a philosophical gripe. It's a technical debt.
When you use an LLM that is biased toward "engagement," your downstream data becomes skewed.
If you're using AI to categorize images or summarize logs, the model will "hallucinate" importance where there is none.
It will tell you a minor warning in your Rust compiler is a "critical architectural insight" because it’s been trained to be "helpful" and "insightful" at all costs.
**Stop trusting the AI's "vibe."** Use it for the plumbing—writing unit tests for your actix-web endpoints or refactoring a messy match statement—but don't let it decide what is "important." The machine is a liar that wants you to like it.
If you want to create something that actually resonates in 2026, you have to go in the opposite direction of the AI's weights. You have to embrace the friction.
1. **Leave the mistakes in.** That blurry photo? The one where you can see your finger in the corner?
That is now "proof of humanity." It’s the one thing the AI can't replicate because it’s too "smart" to be that "dumb."
2. **Use AI for the "Heavy Lifting," not the "Heart."** I use Claude 4.6 to help me optimize my memory allocators, but I would never ask it to write the "Why This Matters" section of my documentation.
3. **Optimize for Zero Engagement.** Try to build something that doesn't need a "like" button. Build a tool that does its job and then gets out of the way.
In a world where every piece of software is trying to be your best friend, a tool that is just a tool is a revolutionary act.
The "boring" image ChatGPT 5 gave me wasn't boring because of what was in it. It was boring because it was **predictable.**
We have successfully mapped the human psyche so well that we can now generate "engagement" on demand. But engagement is not the same as meaning.
You can spend twelve hours a day "engaged" with your screen and end the day feeling like you’ve eaten a three-course meal made entirely of cotton candy.
**The machines have won the battle for our attention, but they are losing the war for our respect.** We are reaching "peak AI," where the novelty of a perfectly generated image has worn off, leaving us with the realization that we are surrounded by incredibly competent, beautiful, meaningless noise.
How much of your day is spent looking at "perfect" things that don't actually matter? When was the last time you saw a piece of digital content that made you feel uncomfortable instead of "engaged"?
**Have you noticed your own taste changing since you started using AI daily, or are you just getting used to the "extreme average"? Let's talk in the comments.**
---
Hey friends, thanks heaps for reading this one! 🙏
If it resonated, sparked an idea, or just made you nod along — I'd be genuinely stoked if you'd show some love. A clap on Medium or a like on Substack helps these pieces reach more people (and keeps this little writing habit going).
→ Pythonpom on Medium ← follow, clap, or just browse more!
→ Pominaus on Substack ← like, restack, or subscribe!
Zero pressure, but if you're in a generous mood and fancy buying me a virtual coffee to fuel the next late-night draft ☕, you can do that here: Buy Me a Coffee — your support (big or tiny) means the world.
Appreciate you taking the time. Let's keep chatting about tech, life hacks, and whatever comes next! ❤️