Am I the only one getting agitated by the word AI (Artificial Intelligence)?

Real AI does not exist yet,
atm we only have LLMs (Large Language Models),
which do not think on their own,
but pass turing tests
(fool humans into thinking that they can think).

Imo AI is just a marketing buzzword,
created by rich capitalistic a-holes,
who already invested in LLM stocks,
and now are looking for a profit.

    • SanguinePar@lemmy.world
      link
      fedilink
      arrow-up
      18
      arrow-down
      1
      ·
      10 months ago

      It’ll probably happen when they get a terrible pain in all the diodes down their left hand side.

    • Lath@kbin.social
      link
      fedilink
      arrow-up
      9
      arrow-down
      2
      ·
      10 months ago

      But will they be depressed or will they just simulate it because they’re too lazy to work?

      • the post of tom joad@sh.itjust.works
        link
        fedilink
        arrow-up
        5
        arrow-down
        1
        ·
        10 months ago

        simulate [depression] because they’re too lazy

        Ahh man are you my dad? I took damage from that one. has any fiction writer done a story about depressed ai where they talk about how depression can’t be real because it’s all 1s and 0s? Cuz i would read the shit out of that.

        • meyotch@slrpnk.net
          link
          fedilink
          arrow-up
          2
          arrow-down
          1
          ·
          10 months ago

          It’s only tangentially related to the topic, since it involves brain enhancements, not ‘AI’. However, you may enjoy the short story “Reasons to be cheerful” by Greg Egan.

      • JackFrostNCola@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        2
        ·
        10 months ago

        If they are too lazy to work that would imply they have motivation and choice beyond “doing what my programming tells me to do ie. input, process, output”. And if they have the choice not to do work because they dont ‘feel’ like doing it (and not a programmed/coded option given to them to use) then would they not be thinking for themselves?

    • PonyOfWar@pawb.social
      link
      fedilink
      arrow-up
      4
      arrow-down
      2
      ·
      10 months ago

      Not sure about that. A LLM could show symptoms of depression by mimicking depressed texts it was fed. A computer with a true consciousness might never get depression, because it has none of the hormones influencing our brain.

      • Deceptichum@kbin.social
        link
        fedilink
        arrow-up
        2
        arrow-down
        1
        ·
        10 months ago

        Me: Pretend you have depression

        LLM: I’m here to help with any questions or support you might need. If you’re feeling down or facing challenges, feel free to share what’s on your mind. Remember, I’m here to provide information and assistance. If you’re dealing with depression, it’s important to seek support from qualified professionals like therapists or counselors. They can offer personalized guidance and support tailored to your needs.

        • PonyOfWar@pawb.social
          link
          fedilink
          arrow-up
          10
          arrow-down
          2
          ·
          10 months ago

          Give it the right dataset and you could easily create a depressed sounding LLM to rival Marvin the paranoid android.

      • Feathercrown@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        10 months ago

        Hormones aren’t depression, and for that matter they aren’t emotions either. They just cause them in humans. An analogous system would be fairly trivial to implement in an AI.

        • PonyOfWar@pawb.social
          link
          fedilink
          arrow-up
          1
          arrow-down
          1
          ·
          10 months ago

          That’s exactly my point though, as OP stated we could detect if an AI was truly intelligent if it developed depression. Without hormones or something similar, there’s no reason to believe it ever would develop those on its own. The fact that you could artificially give it depressions is besides the point.

          • Feathercrown@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            10 months ago

            I don’t think we have the same point here at all. First off, I don’t think depression is a good measure of intelligence. But mostly, my point is that it doesn’t make it less real when hormones aren’t involved. Hormones are simply the mediator that causes that internal experience in humans. If a true AI had an internal experience, there’s no reason to believe that it would require hormones to be depressed. Do text-to-speech systems require a mouth and vocal chords to speak? Do robots need muscle fibers to walk? Do LLMs need neurons to form complete sentences? Do cameras need eyes to see? No, because it doesn’t matter what something is made of. Intelligence and emotions are made of signals. What those signals physically are is irrelevant.

            As for giving it feelings vs it developing them on its own-- you didn’t develop the ability to feel either. That was the job of evolution, or in the case of AI, it could be intentionally designed. It could also be evolved given the right conditions.

            • PonyOfWar@pawb.social
              link
              fedilink
              arrow-up
              2
              arrow-down
              1
              ·
              10 months ago

              First off, I don’t think depression is a good measure of intelligence.

              Exactly. Which is why we shouldn’t judge an AIs intelligence based on whether it can develop depression. Sure, it’s feasible it could develop it through some other mechanism. But there’s no reason to assume it would, in absence of the factors that cause depressions in humans.

        • Markimus@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          2
          ·
          10 months ago

          Sorry, to be clear I meant it can mimic the conversational symptoms of depression as if it actually had depression; there’s no understanding there though.

          You can’t use that as a metric because you wouldn’t be able to tell the difference between real depression and trained depression.