I’m rather curious to see how the EU’s privacy laws are going to handle this.

(Original article is from Fortune, but Yahoo Finance doesn’t have a paywall)

  • Veraticus@lib.lgbt
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    1
    ·
    1 year ago

    I never said the brain (or memory) was a database. I said it was more like a database than what LLMs have, which is nothing.

    • SpiderShoeCult
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      And human beings are more like a fungus (eukaryotes, saprophites) than an LLM is, that doesn’t mean we’re mushrooms.

      However, the human brain is more like an LLM than a database, because the LLM was modelled after the human brain. It’s also very similar in the way that nobody actually can tell precisely how it works, for some reason it just does.

      Now I wouldn’t worry about philosophical implications about the nature of consciousness and such, we’re a long way and we’ll find a way of screwing it up.

      I do question why people are so vehement to always point out what we ‘have’ and how special we are. Nobody sane is saying LLMs are human consciousness 2.0. So why act threatened?

      • Veraticus@lib.lgbt
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        1 year ago

        Lol what the fuck? We know exactly how LLMs work. It’s not magic, and it’s nothing like a human brain. They’re literally word frequency algorithms. There’s nothing special about them and I’m the opposite of threatened; I think it’s absurd people who patently don’t understand them are weighing on this debate disagreeing with me when it’s obvious their position can best be described as ignorant.

        • SpiderShoeCult
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          I’m just going to leave this here.

          some random article

          A quote from the article, I found especially interesting.

          “As a result, no one on Earth fully understands the inner workings of LLMs. Researchers are working to gain a better understanding, but this is a slow process that will take years—perhaps decades—to complete.”

          Quite an interesting read and I’m sure you can find some others if you want to and try hard enough.

          • Veraticus@lib.lgbt
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            1 year ago

            This is a somewhat sensationalist and frankly uninteresting way to describe neural networks. Obviously it would take years of analysis to understand the weights of each individual node and what they’re accomplishing (if it is even understandable in a way that would make sense to people without very advanced math degrees). But that doesn’t mean we don’t understand the model or what it does. We can and we do.

            You have misunderstood this article if what you took from it is this:

            It’s also very similar in the way that nobody actually can tell precisely how it works, for some reason it just does.

            We do understand how it works – as an overall system. Inspecting the individual nodes is as irrelevant to understanding an LLM as cataloguing trees in a forest tells you the name of the city to which the forest is adjacent.