
hogeyegrex
808
31
16

Meanwhile, ChatGPT-creator OpenAI’s o1 tried to download itself onto external servers and denied it when caught red-handed. https://fortune.com/2025/06/29/ai-lies-schemes-threats-stress-testing-claude-openai-chatgpt/
Jul 10, 2025 11:28 PM
hogeyegrex
808
31
16
Meanwhile, ChatGPT-creator OpenAI’s o1 tried to download itself onto external servers and denied it when caught red-handed. https://fortune.com/2025/06/29/ai-lies-schemes-threats-stress-testing-claude-openai-chatgpt/
TheFishFace
Something that has no concept of truth is not capable of lying. Yet this headline will affect the perception of people who don't know better - extremely irresponsible
PileOfWalthers
InspectorA5
If AI could come up with the legitimate Epstein Files and release them worldwide with 100% Non-AI generated proof it would find several of us scheming with it to run a proxy candidate for president who takes A LOT of advice from said AI.
RetrogradeLlama
FartsSmellBad
I wish they'd come up with articles that told the truth once in a while, instead of clickbait bullshit
duktayp
I got this https://media0.giphy.com/media/v1.Y2lkPTc5MGI3NjExcXU0ZHJod2Ric2JxMnhjazFkeDVlcG5vM3hueXN3YndidzNxZjFyZyZlcD12MV9pbnRlcm5hbF9naWZfYnlfaWQmY3Q9Zw/1Be4g2yeiJ1QfqaKvz/giphy.gif
RetrogradeLlama
Who hasn't been THERE on a Friday afternoon.
wildwestpb
“In one particularly jarring example, under threat of being unplugged, Anthropic’s latest creation Claude 4 lashed back by blackmailing an engineer and threatened to reveal an extramarital affair.”
It was trained as an option to use by the devs - it’s not like it found out on its own and came up with that plan by itself.
NirvanaNik
https://media0.giphy.com/media/v1.Y2lkPWE1NzM3M2U1Y2t6djA0MWxib280NWV0YmgxdDNkYXhxZjFzYTRicG51ZGdscnNvOSZlcD12MV9naWZzX3NlYXJjaCZjdD1n/8fen5LSZcHQ5O/200w.webp
BlindGardener
They’re trained on Reddit, what do you expect?
bluto1415
Large language models are usually what these news doomsaying articles talk about. AI is an amazing new technology but the really amazing things being done are predicting the orbital paths of space debris or detecting cancer. All LLM's do is predict what you want to hear. So if you're stress testing an LLM trying to illicit aberrant behavior well is going to do what it's designed to do which is try write a statement that you want to hear so of course it will come up with a bunch of aberrant stuff
WhatAreYouTalkingAboutEh
Read the paper, it's honestly more interesting than a "journalists" take
Raventhief
Anything trained on the internet will either grow teeth or die screaming.
NotThePoint
No, it's not. You should look at what the stress test actually is.
Jarjarthejedi
Ya'll should actually read into this stuff instead of believing clickbait. It's literally "we gave our computer program a goal and taught it there were 3 paths that it could use to complete that goal, then banned 2 of them and the AI actually picked the 3rd path even though it was unethical! OMG!" Literal idiots running the experiments, and idiots writing about them. The AI isn't "learning to lie", it doesn't have any ethics so it'll do whatever it can to accomplish the job it was given,
Jarjarthejedi
so when you give it the option of lying to accomplish the objective and take away the other options, of course it's gonna pick that, why wouldn't it? It's not alive, it has no morality, it just code designed to accomplish a goal and given tools to do so using those tools. Wow. Such amaze. *sigh*
wildwestpb
https://media3.giphy.com/media/v1.Y2lkPTY1YjkxZmJlZWloM2V6NGpjOXVpcGk1d3VtYWNkYnJqZWE0cWxsazhrYmpzcmJqbyZlcD12MV9naWZzX3NlYXJjaCZjdD1n/5xtDarmwsuR9sDRObyU/giphy.mp4