• @brbposting@sh.itjust.works
      link
      fedilink
      English
      121 month ago

      Can absolutely never blindly trust the hallucinating plagiarism machine.

      It’s great where either facts don’t matter or you’re personally in a position to vet all of its “factual” output 100%. Text revision, prompting for additional perspectives, prompting to challenge beliefs and identify gaps. Reformatting, quick and easy data extraction, outlining, brainstorming.

      • @SpaceNoodle@lemmy.world
        link
        fedilink
        English
        41 month ago

        Reformatting and outlining as long as you go over and revise it again anyway, seemingly making that moot.

        Data extraction as long as you don’t care if the data is mangled.

        Brainstorming is a good one, since off-the-wall ideas can be useful in that context.

        • @sugar_in_your_tea@sh.itjust.works
          link
          fedilink
          English
          51 month ago

          In most cases I’ve seen AI used, the person spends as much time correcting it than they would if they just did the work without AI. So maybe it makes you feel more productive because a bunch of stuff happens all at once, but at least for text generation, I think it’s more of a placebo.

          • Blóðbók
            link
            fedilink
            English
            31 month ago

            It can at least get one unstuck, past an indecision paralysis, or give an outline of an idea. It can also be useful for searching though data.

          • @SpaceNoodle@lemmy.world
            link
            fedilink
            English
            11 month ago

            If all I want is something blatantly false or legible yet nonsensical, like a modern lorem ipsum, it’s a real time-saver.

    • Zos_Kia
      link
      fedilink
      English
      91 month ago

      You cannot in all seriousness use a LLM as a research tool. That is explicitly not what it is useful for. A LLM’s latent space is like a person’s memory : sure there is some accurate data in there, but also a lot of “misremembered” or “misinterpreted” facts, and some bullshit.

      Think of it like a reasoning engine. Provide it some data which you have researched yourself, and ask it to aggregate it, or summarize it, you’ll get some great results. But asking it to “do the research for you” is plain stupid. If you’re going to query a probabilistic machine for accurate information, you’d be better off rolling dice.

      • @SpaceNoodle@lemmy.world
        link
        fedilink
        English
        1
        edit-2
        1 month ago

        Exactly my point - except that the word “reasoning” is far too generous, as it implies that there would be some way for it to guarantee that its logic is sound, not just highly resembling legible text.

        • Zos_Kia
          link
          fedilink
          English
          31 month ago

          I don’t understand. Have you ever worked an office job? Most humans have no way to guarantee their logic is sound yet they are the ones who do all of the reasoning on earth. Why would you have higher standards for a machine?

        • @tsonfeir@lemmy.world
          link
          fedilink
          English
          51 month ago

          Nope, just gotta know what it IS, what it ISN’T, and how to correctly write prompts for it to return data that you can use to formulate your own conclusion.

          When using AI, it’s only as smart as the operator.

              • Zos_Kia
                link
                fedilink
                English
                91 month ago

                No you don’t understand. The word AI, which was invented to describe this kind of technology, should not be used to describe this technology. It should instead be reserved for some imaginary magical technology that may exist in the future.

            • @msage@programming.dev
              link
              fedilink
              English
              430 days ago

              As much as I hate to do this, it is AI, as ML is a part of Artificial Intelligence.

              It isn’t AGI, some might say it may be, but they are wrong. But the model is learning.

                • @msage@programming.dev
                  link
                  fedilink
                  English
                  129 days ago

                  Just the notion of a computer having hallucinations should suggest that it’s doing more than just basic code.

                  It’s not ‘intelligent’, but it has ‘learned’ enough beyond standard CPU instructions.

                  That’s why it’s not a General AI, but it’s still an AI.

                  • @SpaceNoodle@lemmy.world
                    link
                    fedilink
                    English
                    1
                    edit-2
                    29 days ago

                    I also talk about gremlins inside CPUs, but that doesn’t mean I think there are magical critters turning a crank inside them.

                    It’s called a metaphor, brother.

                    Regardless, it’s all code that’s eventually run on a CPU, so there isn’t any step where magic is injected.