• Paradox
    link
    fedilink
    English
    1292 months ago

    Can I download their model and run it on my own hardware? No? Then they’re inferior to deepseek

    • @Teanut@lemmy.world
      link
      fedilink
      English
      492 months ago

      In fairness, unless you have about 800GB of VRAM/HBM you’re not running true Deepseek yet. The smaller models are Llama or Qwen distilled from Deepseek R1.

      I’m really hoping Deepseek releases smaller models that I can fit on a 16GB GPU and try at home.

      • @Padit@feddit.org
        link
        fedilink
        English
        222 months ago

        Well, honestly: I have this kind of computational power at my university, and we are in dire need of a locally hosted LLM for a project, so at least for me as a researcher, its really really cool to have that.

        • @Teanut@lemmy.world
          link
          fedilink
          English
          42 months ago

          Lucky you! I need to check my university’s current GPU power but sadly my thesis won’t be needing that kind of horsepower, so I won’t be able to give it a try unless I pay AWS or someone else for it on my own dime.

      • @Womble@lemmy.world
        link
        fedilink
        English
        52 months ago

        Sure but i can run the decensored quants of those distils on my pc, I dont need to even open the article to know that openai isnt going to allow me to do that and so isnt really relevant.

  • Tony Bark
    link
    fedilink
    English
    222 months ago

    I’d like to see OpenAI compare themselves to other models aside from their own.

  • @Viri4thus@feddit.org
    link
    fedilink
    English
    182 months ago

    I wonder how much this puff piece cost OpenAI? Pretty cheap compared to the damage of being caught with the hand in the proverbial cookie jar.

  • @floquant@lemmy.dbzer0.com
    link
    fedilink
    English
    142 months ago

    Yeah ok we get it, they just release the latest checkpoint of their continuously trained model whenever convenient and make big headlines out of it.

  • katy ✨
    link
    fedilink
    English
    62 months ago

    Someone please write a virus that deletes all knowledge from LLMs.

    • @blind3rdeye@lemm.ee
      link
      fedilink
      English
      9
      edit-2
      2 months ago

      Deleting data from them might not be feasible, but there are other tactics.

      […] trapping AI crawlers and sending them down an “infinite maze” of static files with no exit links, where they “get stuck” and “thrash around” for months, he tells users. Once trapped, the crawlers can be fed gibberish data, aka Markov babble, which is designed to poison AI models.