Recently, there has been considerable interest in large language models: machine learning systems which produce human-like text and dialogue. Applications of these systems have been plagued by persistent inaccuracies in their output; these are often called “AI hallucinations”. We argue that these falsehoods, and the overall activity of large language models, is better understood as bullshit in the sense explored by Frankfurt (On Bullshit, Princeton, 2005): the models are in an important way indifferent to the truth of their outputs. We distinguish two ways in which the models can be said to be bullshitters, and argue that they clearly meet at least one of these definitions. We further argue that describing AI misrepresentations as bullshit is both a more useful and more accurate way of predicting and discussing the behaviour of these systems.

  • @Amoeba_Girl
    link
    English
    215 months ago

    Thank you, that is such a beautiful and liberating vision for the future!

    • @Soyweiser
      link
      English
      165 months ago

      imagine a world with 10^^^ times more bullshit, but all the human bullshiters are unemployed! Able to do what they want (except pay rent).

      • @skillissuer@discuss.tchncs.de
        link
        fedilink
        English
        125 months ago

        i’ll take a world where all ad-makers, middle managers, salesmen, conmen, vcs and people who serve them pptxs filled with good idea powder thonking are unemployed (without the automated salesmen flooding internet tubes with drivel part)