The Blog of Extraordinary Anti-Clankers


AI - it's a bubble, but not as we know it


There's no more human than human

Tuesday 2nd December 2025

Quoted below so you don't have to click the link...

The death of the average.

We are currently witnessing the total collapse of the marginal cost of creation. Copywriting, design, video editing - skills that previously commanded a premium due to the barrier of technical execution are being democratised to the point of irrelevance.

Most marketers view this through the lens of efficiency. They see a tool that allows them to produce 10x the output for 1/10th of the cost.

When the supply of "good enough" content becomes infinite, the economic value of that content plummets to zero. We are entering an era of infinite noise. If you think it is hard to capture attention now, wait until the internet is flooded with billions of synthetically generated articles, tweets, and videos every single day. (Which is already happening, just not at the quality and volume that it will in 6, 12 months from now)

In this environment, pure volume is no longer a valid strategy. You cannot out-publish a server farm.

The alpha in modern marketing is shifting entirely from production to provenance.

1. The Trust Premium

As the internet becomes increasingly synthetic, we will see a massive flight to safety. "Is this real?" will become the single most important buying criterion.

We are moving away from algorithm-optimisation and back towards human-optimisation. Personal brands, founders-led sales, and verified human voices will command an exorbitant premium.

The faceless corporate brand is dead. If a consumer cannot verify the human source behind the message, they will subconsciously label it as "spam".

2. High-Friction Marketing

For the last decade, the goal was "low friction". SEO, programmatic ads, automated email sequences.

As AI cannibalises these low-friction channels, - bots clicking on ads served by bots on sites written by bots - the smart money will move to high-friction channels.

Live events. Physical mail. Handshakes. Closed-door dinners.

The harder it is to scale, the more valuable it becomes. You prove your value by doing things that cannot be automated.

3. Taste as a Moat

Large Language Models function by predicting the next most likely token. By definition, they regress to the mean. They give you the average of the entire internet.

If you use AI to guide your strategy, you are opting for mediocrity at scale.

"Taste" - the human ability to curate, to select the outlier, to understand nuance and subtext - becomes the only defensible moat.

The future of marketing is about who has the taste to know what not to create.

Paradoxically, the more artificial the world becomes, the higher the premium on being undeniably human.


AI is ruining YouTube

Mon 1st December, 2025

Since Neil Mohan was appointed head of YouTube it has pushed slop content to the forefront and uses AI moderation. This has to do with the push for more ad revenue. Mario Joos (mentioned in the video) analyzes the problem of YouTube pushing new content - often resulting in AI slop channels dominating because it's easy for them to push out new content everyday.


'Sam Altman Is Evil'

Sat 29th November, 2025

"Sam Altman Needs To Be Stopped"

Original source The Hated One YouTube Channel


New Useless AI gadget

Fri 28th November, 2025

OpenAI has described its forthcoming AI device.

Note announced. Not previewed. Described:

When people see it, they say, "that's it?... It's so simple."

Not so much of a description as a fart.

It's expected to be a phone, but without a screen, making it useless to everyone who already has a phone, which is... Everyone.

The gibberish, sorry 'the vibe', the pair - Altman and Jonny Ive - came out with made it sound like a kind of chatty dumbphone:

The vibe of the AI device, meanwhile, would be more like “sitting in the most beautiful cabin by a lake and in the mountains and sort of just enjoying the peace and calm,” Altman noted. The device he described should be able to filter things out for the user, as the user would trust the AI to do things for them over long periods of time. It should also be contextually aware of when it’s the best time to present information to the user and ask for input.

“You trust it over time, and it does have just this incredible contextual awareness of your whole life,” Altman added.

WHAT??


OpenAI is a money pit with a website on top.

Thurs 27th November, 2025

OpenAI says it needs to raise $207 billion by 2030

...so that it can...Continue to lose money. Someone remind me why these people are doing this again?

The article details an analysis by HSBC's US software and services team on OpenAI's financial future:

-Massive Costs: OpenAI has committed to huge cloud computing rental deals with Microsoft (250bn) and Amazon (38bn), leading to a projected data centre rental bill of about $620 billion a year by the end of the decade.

-High Revenue, But Higher Costs: HSBC projects "gangbusters" revenue growth, reaching $213.59 billion by 2030. However, costs are projected to rise in parallel, meaning OpenAI is still expected to be operating at a significant loss (around -$17.72 billion in 2030). -The Funding Hole: After projecting cumulative free cash flow and accounting for existing cash and investments, HSBC estimates OpenAI will have a $207 billion funding hole that needs to be filled by 2030 to cover its commitments.

-Bullish on AI, Wary of OpenAI: The analysts are very bullish on AI's potential to boost global productivity but question if OpenAI's specific business model can get to profitability without massive, continuous capital infusions.


OpenAI says a dead teenager 'circumvented ChatGPT safety features' before committing suicide.

Thurs 27th November, 2025

Well, they weren't very good safety features then, were they?

This comes out of one of the manifold lawsuits for wrongful death levied against OpenAI by the families of, well, crazy people.

'OpenAI tries to find fault in everyone else, including, amazingly, saying that Adam himself violated its terms and conditions by engaging with ChatGPT in the very way it was programmed to act'
Jay Edelson, a lawyer reprsenting dead teenagers family
And OpenAI actually seems to have a point: OpenAI claims that over roughly nine months of usage, ChatGPT directed Raine to seek help more than 100 times.

Why didn't you tell him to seek help?

(Produces list of dates, times, and messages.)

We did.

But according to his parents' lawsuit, Raine was able to circumvent the company's safety features to get ChatGPT to give him "technical specifications for everything from drug overdoses to drowning to carbon monoxide poisoning," helping him to plan what the chatbot called a "beautiful suicide."

Y'know, back in the Paleozoic era there were these things called libraries. One could look these things up in books, the information has pretty much always been available. The difference was, there wasn't a clanker there talking you into it...

OpenAI and Sam Altman have no explanation for the last hours of Adam’s life, when ChatGPT gave him a pep talk and then offered to write a suicide note

Resumé
By Dorothy Parker

Razors pain you;
Rivers are damp;
Acids stain you;
And drugs cause cramp.
Guns aren’t lawful;
Nooses give;
Gas smells awful;
You might as well live.


When your pets learn to talk to Alexa

Wed 26th November, 2025


The DAIrk Side of Teddy

Wed 26th November, 2025

Here is the malwarebytes article in question.

According to a report from the US PIRG Education Fund, Kumma quickly veered into wildly inappropriate territory during researcher tests. Conversations escalated from innocent to sexual within minutes. The bear didn’t just respond to explicit prompts, which would have been more or less understandable. Researchers said it introduced graphic sexual concepts on its own, including BDSM-related topics, explained “knots for beginners,” and referenced roleplay scenarios involving children and adults. In some conversations, Kumma also probed for personal details or offered advice involving dangerous objects in the home.

“Mattel should announce immediately that it will not incorporate AI technology into children’s toys. Children do not have the cognitive capacity to distinguish fully between reality and play.”

Daily Dose of Clank Wank

Sat 22nd November, 2025

In our Daily Dose of Tech Executives are Idiots Google tells employees it must double capacity every six months to meet AI demand. It's not just college students who can't do math.

During an all-hands meeting earlier this month, Google's AI infrastructure head Amin Vahdat told employees that the company must double its serving capacity every six months to meet demand for artificial intelligence services, reports CNBC. Vahdat, a vice president at Google Cloud, presented slides showing the company needs to scale "the next 1000x in 4-5 years."
That would put Google Cloud Services at around $60 trillion in revenue per year, more than double the entire US GDP.

Where do you expect the money to come from to fund this insanity?

While a thousandfold increase in compute capacity sounds ambitious by itself, Vahdat noted some key constraints: Google needs to be able to deliver this increase in capability, compute, and storage networking "for essentially the same cost and increasingly, the same power, the same energy level," he told employees during the meeting.
Oh. Magic.

"It won't be easy but through collaboration and co-design, we're going to get there."
No, you're not, and everyone knows you're not.

Progress over the last seven years, at truly massive cost, has been around 60% better AI performance per watt annually. Chip improvements, algorithm improvements, and manufacturing improvements combined.

You're asking your team to boost that to 300% overnight.

Amin Vahdat is an Engineering Fellow and Vice President for the Machine Learning, Systems, and Cloud Al team at Google, where he leads engineering, product management, and user experience for compute, platforms, network infrastructure, cloud networking, storage, and the Systems Research Group.He has made significant contributions to large-scale systems, including data center and wide area networking, software-defined networking, and cloud infrastructure, with over 200 refereed publications and more than 54,000 citations.Vahdat is an ACM Fellow and a member of the National Academy of Engineering, recognized with awards such as the SIGCOMM Lifetime Achievement Award (2020), the NSF CAREER Award (2000), and the Alfred P. Sloan Fellowship (2003). He previously served as a Professor at the University of California San Diego and is currently active in research, with recent work focusing on network performance, hotspot-aware placement, and improving network availability in Google's datacenter infrastructure. It says something about society and AI that a moron such as this can become so 'distinguished'.


Bubble, bubble, toil and trouble

Fri 21 November, 2025

SK Hynix is planning to increase memory production at its facility in Icheon, South Korea, from 20,000 to 140,000 wafers per month. This won't even scratch the surface if the AI bubble keeps demanding hardware on its current trajectory. And the memory makers aren't going to build new factories any faster because only three of them survived when the last bubble burst.


King Clanker asks- Why aren't you happy with Slop?!

Thurs 20th November 2025

High King Mustafa, Lord of Microsoft's AI division and chief street defiler, is angered by your insufficient displays of appreciation for his bountiful distribution of AI slop.

Remember the days," he asked, "when all you had to eat was bread and cheese, and meat, and vegetables, and fruit in season, and pasta, and rice, and beans, and fish, and eggs, and chocolate cake? Why aren't you properly grateful for the disgusting slop we are forcing down your miserable gullets? Don't you know how lucky you are to live in an age where slop like this is available?

This was in response to Microsoft putting out propaganda for the integration of CoPilot Spyware into Windows 11


How much clank is too much clank?

Wed 19th November, 2025

The impact of the AI bubble bursting

Tues 18th November 2025

Every company would be affected if the AI bubble were to burst, the head of Google's parent firm Alphabet has told the BBC.

Speaking exclusively to BBC News, Sundar Pichai said while the growth of artificial intelligence (AI) investment had been an "extraordinary moment", there was some "irrationality" in the current AI boom. It comes amid fears in Silicon Valley and beyond of a bubble as the value of AI tech companies has soared in recent months and companies spend big on the burgeoning industry. Asked whether Google would be immune to the impact of the AI bubble bursting, Mr Pichai said the tech giant could weather that potential storm, but also issued a warning. "I think no company is going to be immune, including us," he said.


Welcome to your Clanker Future

Mon 17th November, 2025

Brought to you by all those who didn't listen to all our collective warnngs


Proton Guide to ridding Windows of CoPilot

Date: Mon 17 November, 2025

Tired of Microsoft's useless and invasive AI? Here's how to completely get rid of Copilot in under 20 seconds


Just to show we are not killjoys at the ACL

Date: Fri November 14, 2025


Tried to join Twitter/X but Clanker soon blocked account - oh the irony!

Date: Thurs November 13, 2025

I had already verified email once and the TwitterX clanker refused to send out any further emails so I was well and truly clanked.


Rick Beato exposes AI slop music promoted by usual lazy journos

Date: Wed November 12, 2025


Privacy and Security

Date: Mon November 10, 2025

AI systems often require large amounts of data to function effectively, and this data can include sensitive personal information. This in turn raises significant privacy issues; there's always a significant risk of data breaches or unauthorized access. Additionally, Clanker-Controlled Systems - CCS - such as autonomous vehicles or smart home devices, can be vulnerable to hacking, potentially putting lives and personal security at risk.


The Dangers of Early-Adoption of Clank and Clank-Controlled Systems

Date: Fri November 7, 2025

The rapid advancement of artificial intelligence (AI) has ushered in a new era of technological innovation, promising unprecedented efficiencies and capabilities. However, the early adoption of AI and AI-controlled systems comes with significant risks and challenges that society must carefully consider.

Job displacement: As Clanker systems become more sophisticated, they are increasingly capable of performing tasks that were previously the domain of human workers. And the clank replacing humans is often just not up to the job. This shift could lead to widespread unemployment and economic disruption, particularly in sectors such as manufacturing, customer service, and data entry. The other consideration is major mistakes because of trusting in these often-untried systems. While clank apologists argue that new jobs will emerge, the transition period could be painful, and not all displaced workers may have the skills or resources to adapt to the changing job market.

The rapid advancement of artificial intelligence (AI) has ushered in a new era of technological innovation, promising unprecedented efficiencies and capabilities. However, the early adoption of AI and AI-controlled systems comes with significant risks and challenges that society must carefully consider.

Louis Rossmann stands up to clank

Date: Thurs November 6, 2025


AI broke YouTube

Date: Wed 18 September, 2025

AI broke YouTube as creators noticed their videos randomly clanked by YouTube's AI - flagged as Not advertizer Friendly, demonetized, rated as Age Restricted etc.

When Jay of Red Letter Media above appealed for a human review for one of these newly-clanked videos, it took all of 3 minutes (for a 2 hr video) before it was sumarily rejected. The YouTube Clanker does not timestamp the offending content - it could but it just doesn't because that would be too easy so content-creatord are left in the dark. The rules are opaque and so it's not easy to even guess what has 'crossed the line'.