• AggressivelyPassive@feddit.de
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 year ago

    I’m surprised, these models don’t have something like a “ground truth layer” by now.

    Given that ChatGPT for example is completely unspecialized, I would have expected that relatively there’s a way to hand encode axiomatic knowledge. Like specialized domain knowledge or even just basic math. Even tieried data (i.e. more/less trusted sources) seem not to be part of the design.

    • Splodge5@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      Because it’s not designed to be a knowledge base, it’s designed to imitate human communication. It’s the same reason why ChatGPT can’t do maths - it doesn’t “know” anything, it just predicts the most likely word/bit-of-a-word to come next. ChatGPT being as good as it is at, say, writing code given a natural language prompt is sort of just a happy accident, but people now expect that to be it’s primary function.

    • Drewelite
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      I think this is something that’s easier said than done. Maybe at our current level, but as these AI get more advanced… What is truth? Sure mathematics seems like an easy target until we consider one of the best use cases for AI could be theory. An AI could have a fresh take on our interpretation of mathematics, where these base level assumptions would actually be a hindrance.

      • AggressivelyPassive@feddit.de
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        I mean, let’s be honest here: AI will not be primarily used to find out new truthiness about the universe, but order butter at the right time. Or write basic essays, code, explain known things.

        That kind of knowledge could easily be at categorized.