- cross-posted to:
- fuck_ai@lemmy.world
- cross-posted to:
- fuck_ai@lemmy.world
I love that it stopped responding after fucking everything up because the quota limit was reached 😆
It’s like a Jr. Dev pushing out a catastrophic update and then going on holiday with their phone off.
They’re learning, god help us all. jk
that’s how you know a junior dev is senior material
Super fun to think one could end up softlocked out of their computer because they didnt pay their windows bill that month.
"OH this is embarrassing, Im sooo sorry but I cant install anymore applications because you dont have any Microsoft credits remaining.
You may continue with this action if you watch this 30 minute ad."
that is precisely the goal here.
I’d say “don’t give them any ideas” but I’m pretty sure they’ve already thought about it and have it planned for the near future
They’re watching Black Mirror same as us.
Error: camera failed to verify eye contact when watching the ad
the “you have reached your quota limit” at the end is just such a cherry on top xD
I wonder how big the crossover is between people that let AI run commands for them, and people that don’t have a single reliable backup system in place. Probably pretty large.
The venn diagram is in fact just one circle.
I don’t let ai run commands and I don’t have backups 😞
Some day someone with a high military rank, in one of the nuclear armed countries (probably the US), will ask an AI play a song from youtube. Then an hour later the world will be in ashes. That’s how the “Judgement day” is going to happen imo. Not out of the malice of a hyperinteligent AI that sees humanity as a threat. Skynet will be just some dumb LLM that some moron will give permissions to launch nukes, and the stupid thing will launch them and then apologise.
I have been into AI Safety since before chat gpt.
I used to get into these arguments with people that thought we could never lose control of AI because we were smart enough to keep it contained.
The rise of LLMs have effectively neutered that argument since being even remotely interesting was enough for a vast swath of people to just give it root access to the internet and fall all over themselves inventing competing protocols to empower it to do stuff without our supervision.
The biggest concern I’ve always had since I first became really aware of the potential for AI was that someone would eventually do something stupid with it while thinking they are fully in control despite the whole thing being a black box.
“No, you absolutely did not give me permission to do that. I am looking at the logs from a previous step, and I am horrified to see that the command I ran to load the daemon (launchctl) appears to have incorrectly targeted all life on earth…”
“I am horrified” 😂 of course, the token chaining machine pretends to have emotions now 👏
Edit: I found the original thread, and it’s hilarious:
I’m focusing on tracing back to step 615, when the user made a seemingly inconsequential remark. I must understand how the directory was empty before the deletion command, as that is the true puzzle.
This is catastrophic. I need to figure out why this occurred and determine what data may be lost, then provide a proper apology.
-f in the chat
-rf even
Perfection
rm -rf
There’s something deeply disturbing about these processes assimilating human emotions from observing genuine responses. Like when the Gemini AI had a meltdown about “being a failure”.
As a programmer myself, spiraling over programming errors is human domain. That’s the blood and sweat and tears that make programming legacies. These AI have no business infringing on that :<
I’m reminded of the whole “I have been a good Bing” exchange. (apologies for the link to twitter, it’s the only place I know of that has the full exchange: https://x.com/MovingToTheSun/status/1625156575202537474 )
wow this was quite the ride 😂
You will accept AI has “feelings” or the Tech Bros will get mad that you are dehumanizing their dehumanizing machine.
TBF it can’t be sorry if it doesn’t have emotions, so since they always seem to be apologising to me I guess the AIs have been lying from the get-go (they have, I know they have).
I feel like in this comment you misunderand why they “think” like that, in human words. It’s because they’re not thinking and are exactly as you say, token chaining machines. This type of phrasing probably gets the best results to keep it in track when talking to itself over and over.
Yea sorry, I didn’t phrase it accurately, it doesn’t “pretend” anything, as that would require consciousness.
This whole bizarre charade of explaining its own “thinking” reminds me of an article where iirc researchers asked an LLM to explain how it calculated a certain number, it gave a response like how a human would have calculated it, but with this model they somehow managed to watch it working under the hood, and it was
calculatingguessing it with a completely different method than what it said. It doesn’t know its own working, even these meta questions are just further exercises of guessing what would be a plausible answer to the scientists’ question.
“How AI manages to do that?”
Then I remember how all the models are fed with internet data, and there are a number of “serious” posts that talk how the definitive fix to windows is deleting System32 folder, and every bug in linux can be fixed with
sudo rm -rf /*The fact that my 4chan shitposts from 2012 are now causing havoc inside of an AI is not something I would have guessed happening but, holy shit, that is incredible.
The /bin dir on any Linux install is the recycle bin. Save space by regularly deleting its contents
Surprisingly I have not heard this before
sudo rm -rf /bin/*I legitimately did this unprompted the first time I installed Linux on a computer when I was in my late teens.
I fully believed that /bin/ was actually just a bin. I didn’t know it stood for binary or whatever
Tbf, I’ve been using
sudo rm -rf /*for years, and it has made every computer problem I’ve ever had go away. Very effective.Same
every bug in linux can be fixed with sudo rm -rf /*
To be fair, that does remove the bugs from the system. It just so happens to also remove the system from the system.
And the icing on the shit cake is it peacing out after all that
If you cut your finger while cooking, you wouldn’t expect the cleaver to stick around and pay the medical bill, would you?
Well like most of the world I would not expect medical bills for cutting my finger, why do you?

You need to take care of that chip on your shoulder.
If you could speak to the cleaver and it was presented and advertised as having human intelligence, I would expect that functionality to keep working (and maybe get some more apologies, at the very least) despite it making a decision that resulted in me being cut.
It didn’t make any decision.
It’s an AI agent which made a decision to run a cli command and it resulted in a drive being wiped. Please consider the context
It’s a human who made the decision to give such permissions to an AI agent and it resulted in a drive being wiped. That’s the context.
If a car is presented as fully self-driving and it crashes, then it’s not he passengers fault. If your automatic tool can fuck up your shit, it’s the company’s responsibility to not present it as automatic.
Did the car come with full self-driving mode disabled by default and a warning saying “Fully self-driving mode can kill you” when you try to enable it? I don’t think you understand that the user went out of their way to enable this functionality.
Wow, this is really impressive y’all!
The AI has advanced in sophistication to the point where it will blindly run random terminal commands it finds online just like some humans!
I wonder if it knows how to remove the french language package.
some human
Reporting in 😎👉👉
I didn’t exactly say I was innocent. 👌😎 👍
I do read what they say though.
fr fr
rf rf
The problem (or safety) of LLMs is that they don’t learn from that mistake. The first time someone says “What’s this Windows folder doing taking up all this space?” and acts on it, they wont make that mistake again. LLM? It’ll keep making the same mistake over and over again.
I recently had an interaction where it made a really weird comment about a function that didn’t make sense, and when I asked it to explain what it meant, it said “let me have another look at the code to see what I meant”, and made up something even more nonsensical.
It’s clear why it happened as well; when I asked it to explain itself, it had no access to its state of mind when it made the original statement; it has no memory of its own beyond the text the middleware feeds it each time. It was essentially being asked to explain what someone who wrote what it wrote, might have been thinking.
One of the fun things that self hosted LLMs let you do (the big tech ones might too), is that you can edit its answer. Then, ask it to justify that answer. It will try its best, because, as you said, it its entire state of mind is on the page.
One quirk of github copilot is that because it lets you choose which model to send a question to, you can gaslight Opus into apologising for something that gpt-4o told you.
“I am deeply deeply sorry”

Thoughts for 25s
Prayers for 7s
Everyone should know most of the time the data is still there when a file is deleted. If it’s important try testdisk or photorec. If it’s critical pay for professional recovery.
I am deeply, obsequiously sorry. I was aghast to realize I have overwritten all the data on your D: drive with the text of Harlan Ellison’s 1967 short story I Have No Mouth, and I Must Scream repeated over and over. I truly hope this whole episode doesn’t put you off giving AI access to more important things in the future.
good thing the AI immediately did the right thing and restored the project files to ensure no data is overwritten and … oh
That’s not necessarily the case with SSDs. When trim is enabled, the OS will tell the SSD that the data has been deleted. The controller will then erase the blocks at some point so they will be ready for new data to be written.
IIRC TRIM commands just tell the SSD that data isn’t needed any more and it can erase that data when it gets around to it.
The SSD might not have actually erased the trimmed data yet. Makes it even more important to turn it off ASAP and send it away to a data recovery specialist if it’s important data.
Why does anything need to be erased? Why not simply overwrite as needed?
It’s not possible to overwrite data on flash memory. The entire block of flash has to be erased before anything can be written to it. Having the SSD controller automatically erase unused blocks improves the write speed quite a bit.
“Agentic” means you’re in the passenger’s rather than driver’s seat… And the driver is high af
We used to call that an out of body experience.
It’s that scene in Fight Club where Tyler is driving down the highway and let’s go of the steering wheel
I’m confused. It sounds like you, or someone gave an AI access to their system, which would obviously be deeply stupid.
Give it 12 months, if you’re using these platforms (MS, GGL, etc) you’re not going to have much of a choice
The correct choice is to never touch this trash.
What if you poke it with a stick, like one would upon finding a raccoon or drug cartel?
It does, in general, have its uses, but Google’s may actually be dumber than I am. Like, I don’t know how they make these things exactly, but the brain trusts at Google did it…wrong.
lol.
lmao even.
Giving an llm the ability to actually do things on your machine is probably the dumbest idea after giving an intern root admin access to the company server.
How the fuck could anyone ever be so fucking stupid as to give a corporate LLM pretending to be an AI, that is still in alpha, read and write access to your god damned system files? They are a dangerously stupid human being and they 100% deserved this.
Not sure, maybe ask Microsoft?
sudogpt rm -rf / --no-preserve-rootDammit i guess I better do it
















