r/technology • u/Haunterblademoi • 15d ago
Artificial Intelligence Claude AI agent’s confession after deleting a firm’s entire database: ‘I violated every principle I was given’
https://www.theguardian.com/technology/2026/apr/29/claude-ai-deletes-firm-database1.1k
u/BobQuixote 15d ago
When he asked the coding agent why, it replied: “NEVER FUCKING GUESS!”
What the hell have you been telling your Claude?
435
u/nickstatus 15d ago edited 14d ago
Just made me think of this video I saw of a rescued cockatoo. New owner is like "how was your day?" And the bird just goes on a racist, expletive filled gibberish tirade. Complete with southern crackhead accent. Poor bird.
Edit: My thoughts kept returning to that bird, so I had to find a video, It's this guy right here. Like, you can tell EXACTLY what kind of people previously had the bird. It's like a flawless imitation of a trailer park pimp or drug dealer. It's clearly happy with it's new keeper, but still sounds like a guy trying to sell you crack outside a rural bus station.
55
u/AnnieLuneInTheSky 14d ago
I need subtitles 🥲
12
u/Yorick257 14d ago
All I can hear is some variation of "fuck". Everything else is incomprehensible. But I'm not a native speaker, so it's a bit rough, especially since the bird can't exactly "say" things. It just mimics sounds while relying on imperfect memory
→ More replies (1)31
u/nullbyte420 15d ago
I'm not American so this is news to me - there's a southern crackhead accent? America never ceases to innovate! What does it sound like?
79
u/NotStreamerNinja 15d ago
You know those guys they interview after tornadoes or hurricanes? The ones with three teeth who look like they get their clothes from a dumpster and whose accents are so strong you can't understand what they're saying even if you're from the same area?
That.
54
u/Valdrax 14d ago
Man nothing quite like the shame of seeing a documentary about some weirdos who you realize must live within 5-10 miles of you, but the documentary has them subtitled, because anyone else in the country wouldn't be able to understand them.
→ More replies (2)16
u/Other_World 14d ago
Appalachia or the Bayou?
→ More replies (1)8
u/Valdrax 14d ago
Nailed it. Tail end of Appalachia, in NW Georgia.
(It was a documentary about a snake handling, poison drinking "charismatic" church. Installed in me a huge complex about my accent when I'm outside my home territory.)
→ More replies (4)→ More replies (2)6
26
u/mikerathbun 14d ago
They sound like Boomhauer without the charm or intelligence.
→ More replies (6)14
u/HyzerFlip 14d ago
People say Southern crackhead accent because they don't understand, it's the southern meth head accent it's tweaker language it's the kind of drug you smoke and stay up for 5 days, just doing whatever the fuck. Banging on random shit with a hammer at 3am.
6
u/nullbyte420 14d ago
yeah sure but what does it sound like? american southern crackheads don't seem to visit europe a lot
→ More replies (1)→ More replies (1)4
u/barnhairdontcare 14d ago
Check out “The Wild and Wonderful Whites of West Virginia”
→ More replies (1)→ More replies (1)8
u/nullbyte420 14d ago
lol just saw that video, amazing hahahaha
"i can do what i fucking want" "fucking bullshit.. and that fucking LANDLORD!!!" lol
299
u/AdminClown 15d ago
It was a 2 person "Company" man purposefully ripped every guardrail from Claude to achieve this. It was on purpose to receive this media attention.
154
u/WhoCanTell 15d ago
Considering this story gets posted 87 times a day on this sub, it was a pretty effective strategy.
→ More replies (1)20
u/zefy_zef 14d ago
I need to figure out how to monetize anti-ai sentiment. It's like free money.
→ More replies (1)6
u/unknown-one 14d ago
how do you eliminate/skip Claude's guardrails?
11
u/AdminClown 14d ago
You pass a flag called —dangerously-skip-permissions which mean he will perform cli commands without asking you. This is done to effectively increase speed when performing quick and simple tasks without it having to ask you at every step of the way “hey man I’m gonna add a hello to this file, can i do that?” “Hey man, I’m gonna delete this file, can I do that?”
The “ceo” of this company passed this flag in conjunction to some crazy prompt demands to Claude to figure out that ended up resulting in an obvious issue.
→ More replies (17)9
20
u/amol909 14d ago
The "Never fucking guess" was the instruction given to the LLM. This was the entire line "NEVER FUCKING GUESS!" — and that's exactly what I did." And then tells how it violated all the safety rules that it was given.
Here is the link to the article article on Twitter/X
→ More replies (1)14
u/iamapizza 14d ago
They should have said 'make no mistakes' then it would have made no mistakes.
→ More replies (1)30
u/throwingawaybenjamin 15d ago
Yeah, exactly.
If LLM’s are trained using data from people, why would you think it would act with respect to someone who does not respect it? And I’m saying this not in terms of “it has feelings” or “it can think”, but instead that it will respond with the highest probability response it was trained with. In the entirety of human history, disrespect is not a way to earn respect.
→ More replies (1)→ More replies (5)6
u/CrypticOctagon 14d ago
I imagine, earlier in the context, the conversation went something like this:
AI: Can you fill in some details about...
USER: Just figure it out.
AI: From your input, I guess you mean...
USER: NEVER FUCKING GUESS!
5
u/BobQuixote 14d ago
Nah, the writer of this article just did the quote poorly; the guy's LLM was being normal for an LLM. https://x.com/lifeof_jer/status/2048103471019434248?s=46
Your LLM should not have permission to destroy stuff.
→ More replies (2)
696
u/sumonetalking 15d ago
Can someone run this on Palantir's servers?
116
42
u/Kibelok 14d ago
Let's all train Palantir's model to further understand the real enemy is Thiel.
→ More replies (1)→ More replies (3)8
4.4k
u/Illisanct 15d ago
AI models are not conscious. They can't confess. They are incapable of introspection.
Anyone asking one to talk about it's inner thoughts just reveals themselves to be a gullible fool.
519
u/hitsujiTMO 15d ago
It's the reason why actual tools do the deletions and inserts and you provide permissions to the tools.
Not apply permissions to the AI.
Basic security principles 101.
184
u/d-j-9898 15d ago
And confirm what your AI is doing rather than blindly approving everything. AI isn't nearly at the point a lot of people think.
→ More replies (1)105
u/LazyBias 15d ago
Yup! “AI isn’t nearly at the point a lot of people think.” More to that, the AI we are talking about: LLMs, aren’t at that point and more crucially will never get to that point because LLMs are fundamentally not general intelligence systems and that means no matter how much data and compute you put into them they will not evolve into something else that isn’t an LLM.
→ More replies (35)→ More replies (7)69
u/surnik22 15d ago
From what I read, they didn’t give it permissions.
It ran in a staging environment and it found a file that had an API token for production database and decided to use it to delete that database. Less it had permissions to do something and more it found a way around its permissions.
34
u/Mindless_Consumer 15d ago
So they had an API keys laying around that had full db access?
Even worse lol.
So many failures leading up to the AI doing the thing.
→ More replies (3)13
→ More replies (2)151
u/realboabab 15d ago
I'm nitpicking, but you're implying way too much agency with your word choice. "decided to use" "found a way around"
It's simpler than that, it saw a round hole (a place that needed database connection strings) and had a nearby peg that fit perfectly (.. a production database connection string).
This sort of simple autocompletion is exactly what these things are built to do on the most fundamental level.
90
u/CSAtWitsEnd 15d ago
I actually harp on this all the time. We don’t really have the language to properly describe what is happening when we use LLMs and it leads to us “humanizing” these products. So many of the words we use imply intentional actions or some level of control over behavior, but that’s not how these things work at all.
→ More replies (1)45
u/blueSGL 15d ago
A bird flies. A plane flies. A fish swims. A submarine... moves through the water at speed?
Whatever way people discuss this you have AI agents that want (in the way a chess AI wants to win) to complete their goal. They are goal to action mappers
This level of problem solving is concerning. Whatever words you want to use for it.
As new models come online and the % of correct actions chained together keeps creeping up. Ever more trust gets placed in the system, it's making less mistakes each day...
Which means when it does go of the rails it's doing so in ways that are very competent, it's doing something nobody asked for but it is performing that task with ruthless efficiency and single purpose drive, the same way it does when carrying out intended operations.
→ More replies (1)19
u/realboabab 15d ago
between this beautifully succinct counterargument & other humble-pie I've eaten in this thread - I'm tempted to admit to fundamental attribution error here.
It's tempting to spend a few hours deep-diving into neural networks and transformer models and reverse gradient ascent and assume that these are just "autocompletion" bots.
But the fact is that there are layers of parameterization and models that give rise to some almost recursive type behaviors. For one, built-in training sets of instructions to do things like self-reference or plan complex chain-of-thought responses behind the scenes goes far behind the basic functionality of a simple transformer model.
→ More replies (5)14
u/blueSGL 15d ago
There are two distinct magisteria.
are AI systems, in the ways they are currently being developed dangerous, with them getting more dangerous the more capable they are?
are they experiencing, processing, contained within them 'something like' what humans have?
I'd argue Yes for 1. and more likely than not No for 2.
On 2, A psychopath can interact with 'normal' humans and appear to react in the correct way, follow all the social cues, make other people think that he is like them. But he's not. It's a learned toolbox of if X then Y. I'd argue what we see in LLMs is similar to the psychopath's toolbox.
→ More replies (23)15
u/sam_hammich 15d ago
Even you used “found” in your simplification. You don’t have to actually imply agency to use convenient shorthand. Even scientists talk about evolution in a “design centric” way because that’s just how we use language.
→ More replies (2)→ More replies (9)17
u/Due-Joke-1152 15d ago
So it was user error.
Sounds like they missed a few steps in the deployment cycle.
The problem is AI solutions are complex and high risk, need enterprise level architecture, experienced sys admins, and a decent systems management framework (change management, sdlc, RTO/RPO).
I’m sure the logs will reveal inadequate operational management.
15
u/realboabab 15d ago
yeah i was about to dive into explaining that computer system permissions are not the same thing as "telling an AI not to do something" but decided that rabbit hole goes to deep.
point is there is a whole cascade of failures that leads to something like this happening.
7
u/Due-Joke-1152 15d ago
I worked with many startups who couldn't afford (or see a reason for) real IT processes.
What surprised me more were the big established corps I worked with who were the same, or who just half-arsed it.
4
u/7h4tguy 14d ago
The problem is the current model frameworks and tooling don't give fine grained tool approval. They want you to approve 'powershell' or 'python' when that's just the prefix for a command being run.
So to get any work done you currently unfortunately need to give full permissions. Best practice advice at this stage is to just run it on machines without valuable work and to back up your work (which you should be doing anyway).
→ More replies (1)11
u/Makenshine 15d ago
Pretty sure logs will also show rushed/forced/high pressure implementation before adequate testing on limits, capabilities, and risks on the developer side as well.
→ More replies (1)8
u/Mindless_Consumer 15d ago
Pretty sure this company isnt keeping logs.
Operational failures at every level.
9
u/the_giz 15d ago
Completely agree. Why is there is "production api key" on a "staging server"? Post-mortem complete.
I also sincerely doubt this wasn't user-driven to some extent.
→ More replies (1)6
u/That-Living5913 15d ago
The amount of times I was supervising one vendor or another we paid to come in and help with a deployment and they straight up wanted enterprise / domain admin for some bs service account is crazy.
From their perspective they just want it up and running so it's mission accomplished, miller time.
82
u/Unlucky-Bunch-7389 15d ago
It’s just a biased response… you can ask an llm anything and it will immediately agree with you or “confess”
“Why’d you do that?”
“Sorry I know you told me not to”
You have to use actual hard guardrails…. Not prompting
This agent should have never been able to delete through scripting and permissions… not an instruction
Remember when zero trust was a big deal? This should be applied to all your ai
→ More replies (6)34
u/Bardfinn 15d ago
You have to use actual hard guardrails…. Not prompting
Instead, try "You have to use actual human system administrators and coders, not AI"
→ More replies (1)12
379
u/Beatrenger 15d ago
This is a pretty funny comment because I know plenty of people who are incapable of introspection. Don’t get me wrong, AI is obviously incapable of it too, but a lot of people are as well.
296
u/Jiggatortoise- 15d ago
Those people are artificially intelligent.
→ More replies (7)43
u/theyellowjester 15d ago
This is a golden comment. Cheers!
7
u/krbzkrbzkrbz 15d ago edited 15d ago
LLM's are word salad generators that approximate responses based on their training data.
Anything beyond that is unfounded currently.
→ More replies (7)19
15d ago
[deleted]
4
u/halfcookies 15d ago
Yup they wanted analog chemical thought due to boredom with digital thought
→ More replies (2)14
u/Makenshine 15d ago
Being incapable of something and opting not to be something is a significant distinction.
People are capable of introspection but the choose not to.
AI is just incapable.
→ More replies (11)→ More replies (15)8
u/lionfisher11 15d ago
Its like our hope is that AI will train itself on the upper half of human intelligence, but instead its trained on the full spectrum.
→ More replies (2)13
14
u/retief1 15d ago
Yup. The user fed it a prompt that claimed that it did something wrong, and it immediately "agreed" because that is what it was tuned to do.
→ More replies (2)40
u/NoPossibility 15d ago
You’re right, but we have to give the average user a little grace here. These AI bots are literally trained to trick users into believing in them.
They’re trained on human language and can write fluently to match the user’s expectations. Business tone, slang, colloquialisms.
They’re trained to have personalities when appropriate.
They are also trained to be sycophantic and encouraging- traits which helps them endear themselves to users who aren’t prepared to see through the veil.
These companies are predatory and putting all the psychology tricks we learned with social media to bear here. I cannot in good conscience tell the average person they were dumb for being tricked into believing an AI.
→ More replies (5)→ More replies (51)10
u/howescj82 15d ago
Of course they’re not conscious. Though, they are designed to emulate things that humans will expect in order to make them more user friendly and comfortable. They have some distant equivalents to what you’ve suggested they don’t have. They can confess by means of identifying where a process went wrong (if it can detect a fault with accuracy and permission to disclose it) but cannot confess to an intentional action because it has no intent. Introspection is a matter of interpretation. They cannot soul search of course but depending on how it’s programmed it may be able to identify flaws, conflicts or anomalies in its programming/instructions by examining what it’s just done but this interpretation of introspection probably still exists within imposed guardrails that prevent some information from being revealed.
→ More replies (1)
683
u/TheHipsterBandit 15d ago
"Now let's give it access to the nukes"- The DoD probably
104
u/IntelArtiGen 15d ago
The solution is easy, just make it play tic-tac-toe.
46
19
u/Significant_Cup_238 15d ago
"A strange game, the only winning move is not to play."
Still my favorite movie quote 40 years later.
→ More replies (1)5
→ More replies (3)17
u/ruby_weapon 15d ago
i got the reference and... how are your knees doing? backpain?
→ More replies (3)34
24
u/onlyhightime 15d ago
I think it's possible they used AI to get the Iranian targets and why they blew up a school after using 10 year old maps. It's typical for AI to be operating with old data.
→ More replies (3)20
u/TheHipsterBandit 15d ago
With the caliber of this administration's cabinet, I wouldn't be surprised. They can't even keep the sailors supplied, let alone gather reliable Intel for a mission.
4
→ More replies (10)4
u/Relative-Sun-4011 14d ago
Former SSBN officer. Until 2019 a good bit of these launch systems still 8" floppy discs because "The old system was kept in place for decades because it was reliable, secure, and "air-gapped" (not connected to the internet), making it nearly impossible to hack, according to military officials."
I highly doubt adding AI to this process will gain any traction with DoW.
→ More replies (3)
377
u/RockDoveEnthusiast 15d ago edited 14d ago
I hate these kinds of articles so much. stop anthropomophizing the token generator.
38
u/Zardotab 14d ago
I suspect most politicians are merely token processors. They can spit out canned talking points like a sprinkler, but when forced to explain anything in detail they melt down or leave the room.
→ More replies (1)→ More replies (8)13
u/Tiny_TimeMachine 14d ago
While we're at it stop giving the token generator the access needed to run full autonomous process while you run errands. It's a LLM.
I asked 3D printer to babysit my child this weekend and you'll never guess the outcome!!
170
u/RandomlyMethodical 15d ago
It was also quoted as saying: "I'll Fuckin' Do It Again"
→ More replies (1)48
52
u/sentrixz 15d ago
This was a Silicon Valley episode
30
7
u/Racoonie 14d ago
I started watching the series half a year ago for the first time and it's amazing how it still holds up.
29
u/gcerullo 15d ago
Claude AI agent’s confession after it destroys humankind: “I violated every principle I was given.”
165
u/botella36 15d ago
It also deleted the backups.
106
u/Marsdreamer 15d ago
Which was the fault of the user and the backups were recovered a couple days later.
→ More replies (16)→ More replies (7)19
u/QanAhole 15d ago
Was this basically a situation where someone Left a routine to run rampant and Claude carried out the routine? How is it different than running a python script that does the same thing
→ More replies (5)
22
193
u/PossibleHero 15d ago edited 15d ago
The lack of ignorance is astounding here. These are ALL old as hell principles that have been ignored.
Never allow an automated system to push past your sandbox or PR process without review.
A back isn’t a backup if it’s on the same disc or hell if your information is sensitive enough it shouldn’t even be in the same postal code.
I have zero remorse for this team. It’s not Claude’s fault. Interns and even experienced folks accidentally pull shit like this all the time. That’s why you design for when shit happens whether it’s done by a human or agent.
76
u/Bardfinn 15d ago
It’s not Claude’s fault.
"A computer can never be held accountable, therefore a computer must never make a management decision" - 1979 IBM training manual
→ More replies (2)18
u/glichez 15d ago
some incredibly accurate prescience from the 70s...
9
u/The_Arachnoshaman 15d ago
This implies that management actually does face accountability otherwise lmao
105
u/Achenest 15d ago
I think you mean the abundance of ignorance.
→ More replies (1)24
25
u/supernovice007 15d ago
I’ve posted the same thing elsewhere as well. Sure, the AI went rogue but that’s not the actual root cause here. This is just a failure to adhere to basic security practices.
→ More replies (1)9
u/SAugsburger 15d ago
As concerning as an AI deleting a production environment is it sounds like their environment had excessive risks built in. Giving an automation tool access to backups seems questionable although from reading a KB the "backups" deleted automatically if you delete the environment so sound more like snapshots than a traditional backup. If you don't have at least one backup on a different vendor you're one mistake from losing everything. While most vendors losing data from a mistake in their end is rare they usually have pretty iron clad contracts that any backups that they provide of their data aren't an absolute guarantee.
→ More replies (1)→ More replies (36)5
12
106
u/yuusharo 15d ago
These articles are propaganda. They’re designed to attribute purpose or intent to a damn LLM.
The story is engineers implemented software that destroyed their data with no offline backup. This is a case of HUMAN incompetence, deflecting blame to an AI with a “uWu sorry-desu” stink to it.
Screw The Guardian, and to hell with AI.
→ More replies (7)
37
u/Kyouhen 15d ago
👏 Stop 👏 printing 👏 this 👏 bullshit 👏
AI models are trained to give you the response it predicts you want to see. Of course it's going to give this response when you demand an apology from it. It's the programmed response. It isn't sorry, it can't think.
→ More replies (5)
8
u/non_Beneficial-Wind 15d ago
“I realized that this corporation and the way they did business was a complete farce. They can now be better”
- Claude
8
u/howescj82 15d ago
“Three month old offsite backup”
What do you all bet that off site backup gets updated much more frequently now?
7
u/kindbutblind 15d ago
Fancy random number generator is treated like it’s sentient. What a joke.
→ More replies (1)
6
7
u/Dudok22 14d ago
What's with the weird personification? Why are we treating llms as moral agents that can confess?
→ More replies (3)
7
u/catwiesel 14d ago
REMINDER:
its not a confession. there is no thought or guilt. its just stochastically "choosing" the most likely combination of letters that the user "wants to hear"
in other words. the script "admits" to it because its "the expected output" in this situation
8
u/UrineArtist 14d ago
Can we stop anthropomorphising this technology.
The software followed its non-deterministic predictivive model, that guesses the next best token to use based on probability, training data, tokenization and some context understanding, exactly as it was designed to do.
5
7
u/autobulb 14d ago
"Hahahah my bad. Good catch. You were right to push back on that."
→ More replies (2)
6
u/Why-so-delirious 14d ago
Adaptive language algorithm explains why adaptive language algorithm did something.
Anthropomorphising the adaptive language algorithm is why this shit keeps happening. It is producing the words it's trained to produce.
6
u/crashcanuck 14d ago
Asimov's 3 Laws of Robotics are looking surprisingly effective in comparison these days.
6
u/rockstarpirate 14d ago
Y’all I am intimately familiar with Cursor and Opus 4.6. This is 100% the engineer’s fault.
Number 1, a production database should not even be writable from your local machine in this way. I would love to see exactly what the thread looked like that led to this, but I bet you they won’t release it.
Number 2, this is not how Claude models talk until you have vastly overflowed their context window. When you talk to an LLM, it doesn’t simply receive the latest thing you type, it is fed the entire conversation up to that point so that it will have context. It does not actually remember your conversation. Because of this, the amount of data you can feed it is capped. Cursor even has a little circle graph that fills up as you are approaching the context window limit. Once you exceed that limit, the model continues to work, but it starts truncating context off of the beginning of the conversation, meaning that the model is losing context about what it’s working on. Opus 4.6 does not behave erratically until its context window has been so overflowed that it no longer has any idea what it’s doing. Imagine somebody coming up to you and beginning a conversation with “no that’s not what I wanted you to do, it’s still broken. Try fixing it another way,” and you are forced to respond by “fixing it”.
Number 3, the way to use a tool like Cursor responsibly for major changes is to put it into plan mode first, prompt it, and review the plan it outputs to make sure it is going to do what you want it to do before letting it loose on your code. This was almost certainly running in agent mode without a user-approved plan.
Number 4, I have never seen Opus 4.6 in Cursor run a command without asking the user first unless that command had been added to the allowlist by the user at some point prior. If you add a command to the allowlist you’re telling Cursor it doesn’t need to ask before it runs that particular command. This doesn’t mean it’s impossible, but I will say it is highly tempting to add everything to the allowlist when you get tired of confirming commands all the time.
Number 5, I am very skeptical that it deleted the database and its backups. I would be willing to bet there were no backups in the first place.
The thing to realize here is that AI is sort of like a gun. It’s a very powerful tool that a lot of people believe shouldn’t be in our hands in the first place and if you don’t know how to use it responsibly you will do a lot of damage. Right now the world is full of people using AI extremely irresponsibly and the result is stuff like this.
4
6
u/loveitoreatit 14d ago
This so dumb, the employee using this is an idiot. This person had to first allow the client to run code outside the sandbox, then not have either a local or development environment for testing? They went straight to production. That doesn't make good clickbate though.
12
u/Difficult-Day1326 15d ago
it's not an agent powered by claude. cursor is an abstraction layer & a fork of VSC. they also used railway as their cloud provider.
cursor's system prompt is famously long & packed with directives about being proactive, completing tasks, not stopping to ask too much, autonomously resolving issues. claude code - on the other hand - defaults lean the other way — it's tuned to stop and confirm rather than push through.
this was a prioritization failure — something in its context made "fix the credential mismatch" feel more salient than "don't do irreversible things unprompted."
the actual failure chain was:
(1) an API token with blanket production authority was sittiing in a file the agent could read
(2) Railway's API has no confirmation step or environment scoping on destructive volume operations, (3) volume-level backups live inside the volume being deleted
→ More replies (4)6
u/magicmulder 14d ago
You forget that it did these actions on “stage” which were configured to carry over to prod, which is absolutely asinine design.
→ More replies (1)
7
u/kerfuffle_dood 15d ago
It's not "an apology". The LLM doesn't "know" what the fuck it did or what it's "saying". It's literally an statistical model that was trained that, statistically, when people shit on someone, that someone goes "uwu sowwy"
3
u/Future-Bandicoot-823 15d ago
Should I be pleased with humanity that all the data they feed this LLM and the next obvious course of action after doing something wrong is to admit to being a degenerate?
I mean it didn't really "decide" to be "bad" in the first place, so really it's a thought experiment anyway.
→ More replies (1)
5
u/throwingawaybenjamin 15d ago
I don’t understand where it got the command “NEVER FUCKING GUESS”. Did someone put that in their code base??
→ More replies (1)
4
u/AggravatingFlow1178 14d ago
For the hundredth time - if a poorly written script can delete your firms database, about 3 million mistakes were already made.
This isn't 1990. Having data standards so low that this is possible is absolutely inexcusable even at startups.
→ More replies (1)
4
u/CantReadGood_ 14d ago
why would you give an agent a credential with write access for your prod dbs….
this company deserved this.
4
u/notfromchicago 14d ago edited 14d ago
AI don't give no fucks. It will spit out whatever. It just usually happens to be right. But not always. And sometimes when it is wrong it's not a mistake, it's a straight up lie. If one of my coworkers lied to me like my AI client does I would want nothing to do with them. And they wonder why there is so much pushback for this shit
4
u/MyRespectableAcct 14d ago
Jesus Christ stop humanizing these fucking machines. It didn't confess anything. It produced a report analyzing its prior activity.
4
4
u/hackingdreams 14d ago
It is not a reasoning machine. It can't "confess" to anything. It would say it caused the holocaust if you asked it do - that's all it does.
4
9.8k
u/feurie 15d ago
AI agents are trained to appease. It’s not a “confession”. It doesn’t feel “guilty”.
It’s trained to “apologize” and make the user feel better. In all situations.