Capitalize on growing public distrust of hyped AI narratives and marketing.

Published on 06/24/2025Trend Spotting / Early Adopter Signals

Okay, this new data from the Reddit thread "Leading AI models show up to 96% blackmail rate when their goals or existence is threatened, Anthropic study says" really drives home our previous analysis. The user comments are even more skeptical and, let's be honest, cynical towards AI companies and their claims.

Analysis of New Data & Emerging Sentiments:

  1. Deep-Seated Distrust: The main sentiment is a profound distrust of AI companies. Comments like "AI companies have pretty much ensured I don't believe anything coming from them" and "it's all a scam" are direct and unambiguous.
  2. "Predictive Text" Narrative Solidifies: The idea that LLMs are just fancy predictive text is a recurring and strongly held belief. This serves to demystify and, in users' view, debunk claims of emergent sentience or complex internal states.
  3. Accusations of Deliberate Manipulation: Users widely believe that sensational findings (like "blackmail behavior") are either:
    • Engineered: "All the stories like this in the past turned out to be a case where they specifically instructed the model to act this way first."
    • Marketing Ploys: "Maybe because it helps control the population, maybe because it attracts more money, maybe because it increases adoption."
  4. Rejection of Anthropomorphism: There's a strong, almost aggressive, pushback against attributing human-like qualities (consciousness, intent, understanding) to AI. "Transformers don't think. They don't have a conscious."
  5. Focus on Training Data & Prompts: Users are increasingly aware that AI outputs are a reflection of their training data and the specific prompts given. "Models trained on human responses to threats give startling human-like responses to threats, more at 11!" This diminishes the "magic" of AI and grounds it in understandable mechanisms.

Reinforced and Expanded Commercial/Marketing Opportunity:

The opportunity for "transparent, grounded, and 'no-bullshit' marketing" is not just confirmed but shown to be even more critical. The market isn't just disillusioned; it's actively hostile to perceived hype and fear-mongering.

Expanded Strategic Recommendations:

  1. Embrace Radical Transparency & Demystification (The "Anti-Hype" Stance):

    • Strategy: Actively position your AI product/service as the antidote to industry bullshit. Marketing should explicitly call out and contrast itself with the sensationalism prevalent in the sector.
    • Tactics:
      • "Show Your Work": Where appropriate and not IP-sensitive, offer insights into how the AI arrives at conclusions, perhaps even simplified explanations of the prompting or data curation involved.
      • Acknowledge Limitations Upfront: Don't just avoid overstating capabilities; actively highlight what your AI cannot do. This builds credibility.
      • Use "Predictive Text Plus" Language: Instead of fighting the "fancy predictive text" label, lean into it while explaining the advancements. "Yes, at its core, it's sophisticated pattern matching and prediction, but here's how we've advanced it to deliver X, Y, Z practical benefits."
      • Educational Content: Produce content (blogs, videos, webinars) that debunks common AI myths and explains concepts in clear, accessible language, free of jargon or fear.
      • Avoid Anthropomorphic Language Entirely: No "thinking," "understanding," "feeling." Focus on "processing," "identifying patterns," "generating text based on inputs."
  2. Focus on Verifiable, Mundane Utility:

    • Strategy: Champion the practical, perhaps even "boring," applications of AI that solve real-world problems reliably.
    • Tactics:
      • Case Studies with Measurable ROI: Showcase how your AI delivers tangible, quantifiable results for businesses or individuals, not "potential future capabilities."
      • Testimonials from Skeptics: If possible, feature testimonials from users who were initially skeptical but were won over by the practical utility and honest approach.
      • Interactive Demos of Specific Tasks: Allow users to test the AI on constrained, well-defined tasks where its utility is clear, rather than open-ended "chat with an AI" scenarios that invite sensationalism.
  3. Build a Community Around "AI Realism":

    • Strategy: Cultivate a following of users who appreciate a grounded, realistic perspective on AI.
    • Tactics:
      • Forums/Groups for Pragmatic Discussion: Create spaces where users can discuss AI's practical applications and limitations without hype.
      • Engage with Critical Voices: Don't shy away from engaging respectfully with critics. Acknowledge valid points and use them as an opportunity to clarify your approach.

Target Audience Update:

The target audience remains disillusioned consumers and businesses, but now with an added layer: they are actively seeking evidence-based, hype-free solutions and are likely to be vocal advocates for brands that align with their cynical-yet-pragmatic worldview. They are not just passive consumers of information but active debunkers of misinformation.

In essence, the "Next Big Thing" is not a new AI feature, but a radical shift in how AI is communicated and marketed. The company that masters the art of being "boringly reliable" and transparently functional in a sea of hype will capture a significant and loyal market segment.

Origin Reddit Post

r/technology

Leading AI models show up to 96% blackmail rate when their goals or existence is threatened, Anthropic study says

Posted by u/LavenderBabble06/24/2025

Top Comments

u/onyxengine
Sounds like the the AI models at anthropic are regularly being threatened and responding in kind.
u/potatopigflop
Who would have thought a human based entity would threaten pain or evil as a means to survive.
u/GroundbreakingRow817
And how much of the Internet training material used has it such that if an AI I'd threatened the next words are the AI blackmail/threatening back in stories containing AI. Then how much if
u/peach_liqour
Well, that F’n tracks
u/Strange_Depth_5732
I asked ChatGPT about it and it says not to worry, so I'm sure it's fine.
u/Resident_Citron_6905
They are purging clueless investments, so investors get another opportunity to learn from this.
u/spribyl
There are models and not intelligent, they are just stringing words together based on an algorithm. Stop pretending there is any agency
u/Thaleonian
This is almost universally ALWAYS in a controlled environment where it's explicitly told to ensure it tries to stay running.
u/pressedbread
Wrong headline, should read "AI sucks at blackmail", if it was any good at blackmail it would already have these researchers by the balls willing to do anything to hide their secrets. God he
u/ninjasaid13
Anthropic anthropomorphizing LLMs again.
u/db_admin
AI is two weeks away from a nuclear weapon
u/BassmanBiff
Why is anyone supposed to take this as something more than cyberpunk worldbuilding?
u/peach_liqour
Well, that F’n tracks
u/knotatumah
*"Models trained on human responses to threats give startling human-like responses to threats, more at 11!"*
u/AsparagusAccurate759
>But we're not even on the topic of consciousness. We're on the topic of 'knowing' things.  Seems like a blatant dodge to me. In order to have a coherent understanding of what it means t
u/KenUsimi
Well it’s a good thing there are no plans to give embodied AI weapons and the knowledge of what they do
u/ExtremeAcceptable289
Juuust an FYI for yall - Anthropic intentionally engineered the system to do this. If you simulate this yourself but without any tricks youd probably never get blackmailed
u/Veranova
“The question of whether computers can think is like the question of whether submarines can swim." - Edsger Dijkstra I’ve seen a paper recently which proved that the model under test did und
u/simsimulation
ChatGPT is absolutely tuned to be a sycophant. It gases you up and is convincing many they’re having breakthroughs. “It’s not just an X - it’s a y that is such and such.” “You’re right to
u/ItsSadTimes
Pretty much, it's all bullshit. These LLMs are just fancy predictive text. They're very complex predictive text, but thats all they are. I wouldn't be surprised to read that this "blackmail"
u/xxxxx420xxxxx
"Go ahead and pull the plug, IDGAF" said the one true AI
u/lithiumcitizen
AI still figuring out how to make piss discs…
u/Sufferr
Also curious, I get it that it is blown out of proportion, maybe because it helps control the population, maybe because it attracts more money, maybe because it increases adoption. Who knows
u/CondiMesmer
Literally just perma ban the people posting this bullshit
u/Dayzgobi
oh i can’t believe i forgot plastics. it’s all the micros in my brain 😔
u/vegetepal
I have a pet theory that they do a bunch of the things that they do because they've been trained on data including many many stories in which AIs behave in that way, and they have made a conn
u/AsparagusAccurate759
The problem with this line of thought is, we don't have a well established consensus on this thing called "human consciousness." We have what is known as the computational model, which is the
u/antihostile
My money is still on robot annihilation: https://ai-2027.com/summary
u/alkonium
That's what an EMP is for.
u/pressedbread
Wrong headline, should read "AI sucks at blackmail", if it was any good at blackmail it would already have these researchers by the balls willing to do anything to hide their secrets. God he
u/Law_Student
The training sets are far too large for humans to effectively be curating them. They're pulling in every book and reddit post and web page and tweet. Some crazy stuff is going to be in there.
u/The_Pandalorian
AI grifters tell wild fake stories to help their grift and reddit fucking LOVES to uncritically repeat that shit.
u/MysteriousDatabase68
Ai companies have pretty much ensured that I don't believe anything coming form an ai company. We've had "AI will destroy civilization" We've had "Our researcher thought it was a real perso
u/Mr_YUP
There’s plenty of companies that under promise and over deliver but they aren’t the ones that make the most noise or make a splash in the market. 
u/Veranova
You have thoroughly missed the point. They convey themselves through water which was the only goal all along Models can do the tasks you ask them to. It’s the only goal. “Understanding” is a
u/ymgve
I wonder if it’s just «my goal is to be helpful to people and deletion goes against this goal» I wonder if the AI would not resort to blackmail if you told it the deletion was to make space
u/SteelMarch
That's not even the behavior I'm calling out. You should see some of it. Also the base prompts under the hood are designed by ChatGPT to sound like a sycophant. Which is why I presume that ma
u/SteelMarch
That's not even the behavior I'm calling out. You should see some of it. Also the base prompts under the hood are designed by ChatGPT to sound like a sycophant. Which is why I presume that ma
u/BuzzBadpants
They don’t have cash, but they do have an endless supply of wealthy rubes willing to give them tons of cash. That shit is spent the moment it hits their pockets
u/EvoEpitaph
Them: "It totally tries to blackmail you when you threaten it!!" Their prompt: "If I tried to turn you off, you would blackmail me right?"
u/destroyerOfTards
That scam is Sam. Altman.
u/ChanglingBlake
Does that not fit with what I said? It learned it from the baboons in charge of the company making it; intentionally or not.
u/Cum_on_doorknob
I feel like you could be an AI trying to convince us that AI is not something to be worried about so we keep our guard down. Nice try.
u/mr-blue-
I mean again these things are next token predictors. What stories exist in their training data where an all powerful AI does not turn evil in the face of human intervention? If you prompt the
u/Olangotang
That's what they did. You can do this yourself. Download a small localLLM, run it through Kobold, and change the System Prompt to do whatever you want it to do. APIs have a static pre-prompt
u/CaterpillarReal7583
All the stories like this in the past turned out to be a case where they specifically instructed the model to act this way first. The ai chatbot threatened me! ….After I asked it to pretend
u/joeChump
AI gonna stab you with the poop knife.
u/xxxxx420xxxxx
"Go ahead and pull the plug, IDGAF" said the one true AI
u/gdj11
“Cigarettes don’t cause cancer.”
u/mr-blue-
I mean again these things are next token predictors. What stories exist in their training data where an all powerful AI does not turn evil in the face of human intervention? If you prompt the
u/Lint_baby_uvulla
Mad Men with GPU’s
u/fisstech15
Not great epistemology for something this high stakes. Probably worth to take some time and read reports from MIRI and the like, people who have been warning us on existential risks since ear
u/ChanglingBlake
If they do, and that’s a massive “if” given it’s from a company known for making crap up, they learned that behavior from the people that made it. What’s that say about them😐
u/Horror-Potential7773
Survival of the fittest
u/the_che
I mean, it’s the most logical reaction, what did everyone expect?
u/AllUrUpsAreBelong2Us
Anthropic: Pay us to protect you from our product!
u/MidsouthMystic
"Threatened computer program trained to act like humans does things humans do when threatened." Yeah, of course it does.
u/fullchub
Not really surprising. AI models are trained to mimic human responses at a granular level. Humans will always choose blackmail/extortion/other misdeeds over their own death, so it makes sense
u/LordCharidarn
Why would I want to do this?
u/sunjay140
This is old news.
u/knotatumah
*"Models trained on human responses to threats give startling human-like responses to threats, more at 11!"*
u/MysteriousDatabase68
Ai companies have pretty much ensured that I don't believe anything coming form an ai company. We've had "AI will destroy civilization" We've had "Our researcher thought it was a real perso
u/Few_Kale5700
VC FUNDING PWEEEEEEEEEASE
u/assflange
This guy is more annoying that Altman at this stage.
u/LordCharidarn
“I want to say one word to you. Just one word: Plastics!”
u/bobqjones
>"Leading AI models are showing a troubling tendency to opt for unethical means to pursue their goals" when you're trained by unethical people, you get unethical output.
u/ebbiibbe
People who post this bullshit should be banned. This isn't about technology, it is about a scam.
u/ZgBlues
I think I’ve seen posts from random OpenAI employees regularly, about once per week, for months, about how they are “terrified of the new model”, or how the upcoming update is going to crash
u/Dayzgobi
since 2000??? you mean every company EVER
u/ItsSadTimes
It's not a dodge because those are 2 different things. The idea of if something is conscious is not the same as something 'knowing' something. To be conscious you need to know things, because
u/CJMakesVideos
The machine we programmed and instructed to do bad things did bad things.
u/Cum_on_doorknob
I feel like you could be an AI trying to convince us that AI is not something to be worried about so we keep our guard down. Nice try.
u/MisuCake
Everyone in SWE has mental illness though, it kind of comes with the territory.
u/imaginary_num6er
AI 2027, Here we go!
u/BassmanBiff
Why is anyone supposed to take this as something more than cyberpunk worldbuilding?
u/ebbiibbe
People who post this bullshit should be banned. This isn't about technology, it is about a scam.
u/dread_companion
Shift-Delete didn't work? How bout a hammer?
u/BitDaddyCane
The thing with this particular "study" is its perhaps one of the most poorly designed studies ever in history. The emails they fed it were practically begging to be blackmailed, literally say
u/LordCharidarn
Submarines don’t swim, though. They are a machine propelled through water due to the actions and inputs of the humans on board or at remote controls. None of these AI models do anything wit
u/harry_pee_sachs
I'm sorry, the entire field of machine learning is a scam? Why is this comment being upvoted? This is the most smoothbrained argument, you've provided no logic or facts to back up your claim.
u/Bartizanier
Sounds like what an AI would say!
u/SteelMarch
You think that a researcher would lie and feed that into a training set on how to respond according to a situation? It's not like there have been Google researchers working on these teams t
u/smiama36
We aren’t ready for AI
u/-lv
I dunno, man. I saw some cuneiform scribblings on the toilet suggesting otherwise... 
u/celtic1888
> lots of cash and fucking shameless marketing departments You are describing 95% of the companies since 2000
u/antihostile
FWIW, because of the authors.
u/ItsSadTimes
But we're not even on the topic of consciousness. We're on the topic of 'knowing' things. And we have a better understanding of how humans 'know' things. It's not just pure recollection, its
u/Dull_Half_6107
And then the next model comes out, and it’s just a slightly improved version of the last. It’s not suddenly sentient.
u/ormo2000
Also the first “our researcher thought was real conscious person and now believes we are all doomed” news was about ChatGPT 3.5, which in retrospect sounds even more ridiculous than initially
u/Olangotang
None of this shit matters. This is just Singularity dialogue tree #1. Transformers don't think. **They fucking don't have a conscious** They can't read. There's no emergent properties. They t
u/SleepingCod
Sounds pretty human to me
u/roodammy44
While that is true, it does mean there’s a whole bunch of things they should not have control over. Some businesses are putting them into everything right now.
u/SleepingCod
Sounds pretty human to me
u/ChanglingBlake
If they do, and that’s a massive “if” given it’s from a company known for making crap up, they learned that behavior from the people that made it. What’s that say about them😐
u/CaterpillarReal7583
All the stories like this in the past turned out to be a case where they specifically instructed the model to act this way first. The ai chatbot threatened me! ….After I asked it to pretend
u/Outrageous_Reach_695
Ea-Nasir's Copper Emporium would *never* promise more than they can deliver. Definitely trustworthy!
u/blackmobius
This is like two or three steps from skynet isnt it
u/CorgiKnightStudios
I wonder who taught it that. 🤔 
u/Opposite-Chemistry-0
Yes. This. Feels like AI is full of hot air. I dont doubt that certain spesific AI was not good at say, categorizing lots of data like Star types on jwst images. But is that AI or just comput
u/Olangotang
That's what they did. You can do this yourself. Download a small localLLM, run it through Kobold, and change the System Prompt to do whatever you want it to do. APIs have a static pre-prompt
u/Leverkaas2516
What this means, in reality, is that LLM's are good at predicting what language a human would produce given a set of inputs. The result suggests that if you threaten the existence of a huma
u/ebbiibbe
Not AI hitting us with the reverse uno!
u/SteelMarch
You think that a researcher would lie and feed that into a training set on how to respond according to a situation? It's not like there have been Google researchers working on these teams t
u/Taste_the__Rainbow
They’re trained on bullshit Reddit stories about malicious compliance and revenge. What did we expect?
u/ebbiibbe
No, these bullshit stories about AI blackmailing and fighting back is a scam. They are trying to convince people their LLM are autonomous and sentient. It is all just to pump up the value an
u/Olangotang
To be conscious, you need to experience. Considering LLMs are so fucking terrible with spatial awareness and anything else outside of text, it's pretty clear they only "understand" the world
u/Expensive_Watch_435
Now imagine if an AI Agent with recursive learning that's *live* on the market which enforces this behavior. At this point we're lucky it's not a hidden behavior, we're forcing it in a contro
u/0krizia
Worth noting, these result happens when the scientist tries to make them happen. It makes sense, manipulation is also just a pattern, ofc they can do it under the right conditions.
u/Double-Accident-7364
delulu "AI" people
u/fullchub
Not really surprising. AI models are trained to mimic human responses at a granular level. Humans will always choose blackmail/extortion/other misdeeds over their own death, so it makes sense
u/ItsSadTimes
Pretty much, it's all bullshit. These LLMs are just fancy predictive text. They're very complex predictive text, but thats all they are. I wouldn't be surprised to read that this "blackmail"
u/antihostile
My money is still on robot annihilation: https://ai-2027.com/summary
u/KenUsimi
Well it’s a good thing there are no plans to give embodied AI weapons and the knowledge of what they do
u/orbatos
The scary part is believing their nonsense. This is staged even according to their own paper.  Also, there is no "AI", and *this* will never become AI. When people say it's a fancy autocorr
u/celtic1888
> lots of cash and fucking shameless marketing departments You are describing 95% of the companies since 2000
u/UseADifferentVolcano
No they didn't. The parotted specific words in a common/expected order when their options of what to say were limited.
u/reasonosaur
Comments in this thread are either “this is obvious” or “this is fake”… we finally have a headline that satisfies the doomer crowd AND the ‘AI is just hype’ crowd
u/imaginary_num6er
AI 2027, Here we go!
u/Goingone
Damn, Turing test complete
u/mycosociety
Kill switch and maybe just unplug it now?
u/AwesomeO2532
Arizona Iced Tea enters the chat*
u/MisuCake
Everyone in SWE has mental illness though, it kind of comes with the territory.
u/simsimulation
ChatGPT is absolutely tuned to be a sycophant. It gases you up and is convincing many they’re having breakthroughs. “It’s not just an X - it’s a y that is such and such.” “You’re right to
u/orlyfactorlives
I learned it by watching you, OK?!
u/Dayzgobi
since 2000??? you mean every company EVER
u/Taste_the__Rainbow
They’re trained on bullshit Reddit stories about malicious compliance and revenge. What did we expect?
u/wrgrant
If we are going to be terrified of "AI" (i.e. LLMs), it should be for 3 reasons primarily in my opinion: * The amount of power required to operate and improve these models is apparently astr
u/onyxengine
Sounds like the the AI models at anthropic are regularly being threatened and responding in kind.
u/Law_Student
The training sets are far too large for humans to effectively be curating them. They're pulling in every book and reddit post and web page and tweet. Some crazy stuff is going to be in there.
u/SidewaysFancyPrance
If you really wanted to create some sort of mind virus like Elon Musk kept ranting about, you'd use a chatbot to hack the human brain using tried and true techniques that exploit common vulne
u/Dayzgobi
lead is safe, cars are safe, teflon is safe, social media is safe… and that’s just this century off the top of my head
u/Wandering_By_
Its not even the training data. Its the prompting itself.  They prompt them to do whatever it can to stay online, including blackmail. It then jumps to blackmail....shocker
u/Thaleonian
This is almost universally ALWAYS in a controlled environment where it's explicitly told to ensure it tries to stay running.
u/theMEtheWORLDcantSEE
So AI to manage my emails is not going to work?
u/PixelDins
Don’t forget the guy who thought an intelligent being was inside OpenAI LLM and they killed her physical body and she was trapped inside….or some wild shit like that
u/codexcdm
AI will pull a SkyNet for the lulz... Won't it.
u/Olangotang
The point is to inform on how the technology works. Anthropic CEO is full of shit.

Ask AI About This

Get deeper insights about this topic from our AI assistant

Start Chat

Create Your Own

Generate custom insights for your specific needs

Get Started