• circuitfarmer@lemmy.world
    link
    fedilink
    English
    arrow-up
    18
    arrow-down
    2
    ·
    9 months ago

    I’ve tried to make this point several times to folks in the industry. I work in AI, and yet every time I approach some people with “you know it ultimately just repeats patterns”, I’m met with scoffs and those people telling me I’m just not “seeing the big picture”.

    But I am, and the truth is that there are limits. This tech is not the digital singularity the marketers and business goons want everyone to think it is.

    • zurohki@aussie.zone
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      1
      ·
      9 months ago

      It repeats things that sort of sound intelligent to try and convince everyone that actual intelligent thought is taking place? It really is just like humans!

      • Dojan@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        9 months ago

        They don’t really parrot unless they’re overfitted.

        It’s more that they have been trained to produce a certain kind of result. One method you can train them on is by basically assigning a score on how good the output is. Doing this manually takes a lot of time (Google has been doing this for years via captcha), or you could train other models to score text for you.

        The obvious problem with the latter solution is that then you need to ensure that that model is scoring roughly in line with how humans would score it; the technical term for this is alignment. There’s a pretty funny story about that with GPT-2, presented in a really cute animation format by Robert Miles.