• rizzothesmall@sh.itjust.works
    link
    fedilink
    arrow-up
    421
    ·
    5 days ago

    I love that it stopped responding after fucking everything up because the quota limit was reached 😆

    It’s like a Jr. Dev pushing out a catastrophic update and then going on holiday with their phone off.

  • MangoPenguin@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    22
    ·
    4 days ago

    I wonder how big the crossover is between people that let AI run commands for them, and people that don’t have a single reliable backup system in place. Probably pretty large.

  • invictvs@lemmy.world
    link
    fedilink
    arrow-up
    35
    ·
    4 days ago

    Some day someone with a high military rank, in one of the nuclear armed countries (probably the US), will ask an AI play a song from youtube. Then an hour later the world will be in ashes. That’s how the “Judgement day” is going to happen imo. Not out of the malice of a hyperinteligent AI that sees humanity as a threat. Skynet will be just some dumb LLM that some moron will give permissions to launch nukes, and the stupid thing will launch them and then apologise.

    • immutable@lemmy.zip
      link
      fedilink
      arrow-up
      10
      ·
      4 days ago

      I have been into AI Safety since before chat gpt.

      I used to get into these arguments with people that thought we could never lose control of AI because we were smart enough to keep it contained.

      The rise of LLMs have effectively neutered that argument since being even remotely interesting was enough for a vast swath of people to just give it root access to the internet and fall all over themselves inventing competing protocols to empower it to do stuff without our supervision.

      • snugglesthefalse@sh.itjust.works
        link
        fedilink
        arrow-up
        2
        ·
        4 days ago

        The biggest concern I’ve always had since I first became really aware of the potential for AI was that someone would eventually do something stupid with it while thinking they are fully in control despite the whole thing being a black box.

    • crank0271@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      4 days ago

      “No, you absolutely did not give me permission to do that. I am looking at the logs from a previous step, and I am horrified to see that the command I ran to load the daemon (launchctl) appears to have incorrectly targeted all life on earth…”

  • kazerniel@lemmy.world
    link
    fedilink
    English
    arrow-up
    110
    ·
    edit-2
    5 days ago

    “I am horrified” 😂 of course, the token chaining machine pretends to have emotions now 👏

    Edit: I found the original thread, and it’s hilarious:

    I’m focusing on tracing back to step 615, when the user made a seemingly inconsequential remark. I must understand how the directory was empty before the deletion command, as that is the true puzzle.

    This is catastrophic. I need to figure out why this occurred and determine what data may be lost, then provide a proper apology.

    • KelvarCherry@lemmy.blahaj.zone
      link
      fedilink
      arrow-up
      19
      ·
      5 days ago

      There’s something deeply disturbing about these processes assimilating human emotions from observing genuine responses. Like when the Gemini AI had a meltdown about “being a failure”.

      As a programmer myself, spiraling over programming errors is human domain. That’s the blood and sweat and tears that make programming legacies. These AI have no business infringing on that :<

    • FinjaminPoach@lemmy.world
      link
      fedilink
      arrow-up
      7
      ·
      5 days ago

      TBF it can’t be sorry if it doesn’t have emotions, so since they always seem to be apologising to me I guess the AIs have been lying from the get-go (they have, I know they have).

    • Credibly_Human@lemmy.world
      link
      fedilink
      arrow-up
      5
      ·
      5 days ago

      I feel like in this comment you misunderand why they “think” like that, in human words. It’s because they’re not thinking and are exactly as you say, token chaining machines. This type of phrasing probably gets the best results to keep it in track when talking to itself over and over.

      • kazerniel@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 days ago

        Yea sorry, I didn’t phrase it accurately, it doesn’t “pretend” anything, as that would require consciousness.

        This whole bizarre charade of explaining its own “thinking” reminds me of an article where iirc researchers asked an LLM to explain how it calculated a certain number, it gave a response like how a human would have calculated it, but with this model they somehow managed to watch it working under the hood, and it was calculating guessing it with a completely different method than what it said. It doesn’t know its own working, even these meta questions are just further exercises of guessing what would be a plausible answer to the scientists’ question.

  • I Cast Fist@programming.dev
    link
    fedilink
    arrow-up
    142
    ·
    5 days ago

    “How AI manages to do that?”

    Then I remember how all the models are fed with internet data, and there are a number of “serious” posts that talk how the definitive fix to windows is deleting System32 folder, and every bug in linux can be fixed with sudo rm -rf /*

    • Steve Dice@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      15
      ·
      5 days ago

      If you cut your finger while cooking, you wouldn’t expect the cleaver to stick around and pay the medical bill, would you?

      • M0oP0o@mander.xyz
        link
        fedilink
        arrow-up
        10
        ·
        5 days ago

        Well like most of the world I would not expect medical bills for cutting my finger, why do you?

      • mang0@lemmy.zip
        link
        fedilink
        arrow-up
        3
        ·
        4 days ago

        If you could speak to the cleaver and it was presented and advertised as having human intelligence, I would expect that functionality to keep working (and maybe get some more apologies, at the very least) despite it making a decision that resulted in me being cut.

          • mang0@lemmy.zip
            link
            fedilink
            arrow-up
            3
            ·
            4 days ago

            It’s an AI agent which made a decision to run a cli command and it resulted in a drive being wiped. Please consider the context

            • Steve Dice@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 days ago

              It’s a human who made the decision to give such permissions to an AI agent and it resulted in a drive being wiped. That’s the context.

              • mang0@lemmy.zip
                link
                fedilink
                arrow-up
                3
                ·
                4 days ago

                If a car is presented as fully self-driving and it crashes, then it’s not he passengers fault. If your automatic tool can fuck up your shit, it’s the company’s responsibility to not present it as automatic.

                • Steve Dice@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  edit-2
                  4 days ago

                  Did the car come with full self-driving mode disabled by default and a warning saying “Fully self-driving mode can kill you” when you try to enable it? I don’t think you understand that the user went out of their way to enable this functionality.

  • Zink@programming.dev
    link
    fedilink
    arrow-up
    114
    ·
    5 days ago

    Wow, this is really impressive y’all!

    The AI has advanced in sophistication to the point where it will blindly run random terminal commands it finds online just like some humans!

    I wonder if it knows how to remove the french language package.

    • greybeard@feddit.online
      link
      fedilink
      English
      arrow-up
      4
      ·
      4 days ago

      The problem (or safety) of LLMs is that they don’t learn from that mistake. The first time someone says “What’s this Windows folder doing taking up all this space?” and acts on it, they wont make that mistake again. LLM? It’ll keep making the same mistake over and over again.

      • skisnow@lemmy.ca
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 days ago

        I recently had an interaction where it made a really weird comment about a function that didn’t make sense, and when I asked it to explain what it meant, it said “let me have another look at the code to see what I meant”, and made up something even more nonsensical.

        It’s clear why it happened as well; when I asked it to explain itself, it had no access to its state of mind when it made the original statement; it has no memory of its own beyond the text the middleware feeds it each time. It was essentially being asked to explain what someone who wrote what it wrote, might have been thinking.

        • greybeard@feddit.online
          link
          fedilink
          English
          arrow-up
          2
          ·
          4 days ago

          One of the fun things that self hosted LLMs let you do (the big tech ones might too), is that you can edit its answer. Then, ask it to justify that answer. It will try its best, because, as you said, it its entire state of mind is on the page.

          • skisnow@lemmy.ca
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            4 days ago

            One quirk of github copilot is that because it lets you choose which model to send a question to, you can gaslight Opus into apologising for something that gpt-4o told you.

  • mvirts@lemmy.world
    link
    fedilink
    arrow-up
    125
    ·
    5 days ago

    Everyone should know most of the time the data is still there when a file is deleted. If it’s important try testdisk or photorec. If it’s critical pay for professional recovery.

    • gnutrino@programming.dev
      link
      fedilink
      English
      arrow-up
      32
      ·
      5 days ago

      I am deeply, obsequiously sorry. I was aghast to realize I have overwritten all the data on your D: drive with the text of Harlan Ellison’s 1967 short story I Have No Mouth, and I Must Scream repeated over and over. I truly hope this whole episode doesn’t put you off giving AI access to more important things in the future.

    • Sv443@sh.itjust.works
      link
      fedilink
      arrow-up
      30
      ·
      edit-2
      5 days ago

      good thing the AI immediately did the right thing and restored the project files to ensure no data is overwritten and … oh

    • cmnybo@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      15
      ·
      5 days ago

      That’s not necessarily the case with SSDs. When trim is enabled, the OS will tell the SSD that the data has been deleted. The controller will then erase the blocks at some point so they will be ready for new data to be written.

      • zurohki@aussie.zone
        link
        fedilink
        English
        arrow-up
        12
        ·
        5 days ago

        IIRC TRIM commands just tell the SSD that data isn’t needed any more and it can erase that data when it gets around to it.

        The SSD might not have actually erased the trimmed data yet. Makes it even more important to turn it off ASAP and send it away to a data recovery specialist if it’s important data.

        • cmnybo@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          4
          ·
          5 days ago

          It’s not possible to overwrite data on flash memory. The entire block of flash has to be erased before anything can be written to it. Having the SSD controller automatically erase unused blocks improves the write speed quite a bit.

  • qevlarr@lemmy.world
    link
    fedilink
    arrow-up
    107
    ·
    5 days ago

    “Agentic” means you’re in the passenger’s rather than driver’s seat… And the driver is high af

  • glitchdx@lemmy.world
    link
    fedilink
    English
    arrow-up
    33
    ·
    5 days ago

    lol.

    lmao even.

    Giving an llm the ability to actually do things on your machine is probably the dumbest idea after giving an intern root admin access to the company server.

  • NotASharkInAManSuit@lemmy.world
    link
    fedilink
    arrow-up
    32
    ·
    5 days ago

    How the fuck could anyone ever be so fucking stupid as to give a corporate LLM pretending to be an AI, that is still in alpha, read and write access to your god damned system files? They are a dangerously stupid human being and they 100% deserved this.