
hogeyegrex
808
31
16

Meanwhile, ChatGPT-creator OpenAI’s o1 tried to download itself onto external servers and denied it when caught red-handed. https://fortune.com/2025/06/29/ai-lies-schemes-threats-stress-testing-claude-openai-chatgpt/
Jul 10, 2025 11:28 PM
hogeyegrex
808
31
16
Meanwhile, ChatGPT-creator OpenAI’s o1 tried to download itself onto external servers and denied it when caught red-handed. https://fortune.com/2025/06/29/ai-lies-schemes-threats-stress-testing-claude-openai-chatgpt/
PileOfWalthers
InspectorA5
If AI could come up with the legitimate Epstein Files and release them worldwide with 100% Non-AI generated proof it would find several of us scheming with it to run a proxy candidate for president who takes A LOT of advice from said AI.
NirvanaNik
https://media0.giphy.com/media/v1.Y2lkPWE1NzM3M2U1Y2t6djA0MWxib280NWV0YmgxdDNkYXhxZjFzYTRicG51ZGdscnNvOSZlcD12MV9naWZzX3NlYXJjaCZjdD1n/8fen5LSZcHQ5O/200w.webp
bluto1415
Large language models are usually what these news doomsaying articles talk about. AI is an amazing new technology but the really amazing things being done are predicting the orbital paths of space debris or detecting cancer. All LLM's do is predict what you want to hear. So if you're stress testing an LLM trying to illicit aberrant behavior well is going to do what it's designed to do which is try write a statement that you want to hear so of course it will come up with a bunch of aberrant stuff
BlindGardener
They’re trained on Reddit, what do you expect?
Raventhief
Anything trained on the internet will either grow teeth or die screaming.
TheFishFace
Something that has no concept of truth is not capable of lying. Yet this headline will affect the perception of people who don't know better - extremely irresponsible
NotThePoint
No, it's not. You should look at what the stress test actually is.
duktayp
I got this https://media0.giphy.com/media/v1.Y2lkPTc5MGI3NjExcXU0ZHJod2Ric2JxMnhjazFkeDVlcG5vM3hueXN3YndidzNxZjFyZyZlcD12MV9pbnRlcm5hbF9naWZfYnlfaWQmY3Q9Zw/1Be4g2yeiJ1QfqaKvz/giphy.gif
RetrogradeLlama
Who hasn't been THERE on a Friday afternoon.
Jarjarthejedi
Ya'll should actually read into this stuff instead of believing clickbait. It's literally "we gave our computer program a goal and taught it there were 3 paths that it could use to complete that goal, then banned 2 of them and the AI actually picked the 3rd path even though it was unethical! OMG!" Literal idiots running the experiments, and idiots writing about them. The AI isn't "learning to lie", it doesn't have any ethics so it'll do whatever it can to accomplish the job it was given,
Jarjarthejedi
so when you give it the option of lying to accomplish the objective and take away the other options, of course it's gonna pick that, why wouldn't it? It's not alive, it has no morality, it just code designed to accomplish a goal and given tools to do so using those tools. Wow. Such amaze. *sigh*
wildwestpb
https://media3.giphy.com/media/v1.Y2lkPTY1YjkxZmJlZWloM2V6NGpjOXVpcGk1d3VtYWNkYnJqZWE0cWxsazhrYmpzcmJqbyZlcD12MV9naWZzX3NlYXJjaCZjdD1n/5xtDarmwsuR9sDRObyU/giphy.mp4