I have many conversations with people about Large Language Models like ChatGPT and Copilot. The idea that “it makes convincing sentences, but it doesn’t know what it’s talking about” is a difficult concept to convey or wrap your head around. Because the sentences are so convincing.

Any good examples on how to explain this in simple terms?

Edit:some good answers already! I find especially that the emotional barrier is difficult to break. If an AI says something malicious, our brain immediatly jumps to “it has intent”. How can we explain this away?

    • Hucklebee@lemmy.worldOP
      link
      fedilink
      arrow-up
      6
      ·
      7 months ago

      I commented something similair on another post, but this is exactly why I find this phenomenon so hard to describe.

      A teenager in a new group still has some understanding and has a mind. It knows many of the meaning of the words that are said. Sure, some catchphrases might be new, but general topics shouldn’t be too hard to follow.

      This is nothing like genAI. GenAI doesn’t know anything at all. It has (simplified) a list of words that somehow are connected to eachother. But AI has no meaning of a wheel, what round is, what rolling is, what rubber is, what an axle is. NO understanding. Just words that happened to describe all of it. For us humans it is so difficult to understand that something uses language without knowing ANY of the meaning.

      How can we describe this so our brains make sense that you can have language without understanding? The Chinese Room experiment comes close, but is quite complicated to explain as well I think.

      • NeoNachtwaechter@lemmy.world
        link
        fedilink
        arrow-up
        2
        ·
        7 months ago

        How can we describe this so our brains make sense that you can have language without understanding?

        I think it is really impossible to describe in easy and limited words.

      • JackbyDev@programming.dev
        link
        fedilink
        English
        arrow-up
        1
        ·
        7 months ago

        Joel Haver has a sketch in which one person in a group laughs at an inside joke from a trip they didn’t go on. When pressed I think they say something like they laughed because everyone else was. As someone who has been in this situation, it’s true. Even though I don’t understand the specific reference being made, it’s usually being done in a funny manner such that the story telling is enjoyable and humorous. Or I’m able to use context clues to guess what they might be joking about and it’s funny, even if my understanding is off.

      • Turun@feddit.de
        link
        fedilink
        arrow-up
        1
        ·
        7 months ago

        NO understanding. Just words that happened to describe all of it.

        If being able to describe it does not mean understanding, then what is understanding?