Am I the only one getting agitated by the word AI (Artificial Intelligence)?

Real AI does not exist yet,
atm we only have LLMs (Large Language Models),
which do not think on their own,
but pass turing tests
(fool humans into thinking that they can think).

Imo AI is just a marketing buzzword,
created by rich capitalistic a-holes,
who already invested in LLM stocks,
and now are looking for a profit.

  • doctorcrimson@lemmy.world
    link
    fedilink
    arrow-up
    6
    arrow-down
    2
    ·
    edit-2
    11 months ago

    So basically the ability to do things or learn without direction for tasks other than what it was created to do. Example, ChatGPT doesn’t know how to play chess and Deep Blue doesn’t write poetry. Either might be able to approximate correct output if tweaked a bit and trained on thousands, millions, or billions of examples of proper output, but neither are capable of learning to think as a human would.

    • intensely_human@lemm.ee
      link
      fedilink
      arrow-up
      0
      arrow-down
      6
      ·
      11 months ago

      I think it could learn to think as a human does. Humans think by verbalizing at themselves: running their own verbal output back into their head.

      Now don’t get me wrong. I’m envisioning like thousands of prompt-response generations, with many of these LLMs playing specialized roles: generating lists of places to check for X information in its key-value store. The next one’s job is to actually do that. The reason for separation is exhaustion. That output goes to three more. One checks it for errors, and sends it back to the first with errors highlighted to re-generate.

      I think that human thought is more like this big cluster of LLMs all splitting up work and recombining it this way.

      Also, you’d need “dumb”, algorithmic code that did tasks like:

      • compile the last second’s photograph, audio intake, infrared, whatever, and send it to the processing team.

      • Processing team is a bunch of LLMs, each with a different task in its prompt: (1) describe how this affects my power supply, (2) describe how this affects my goal of arriving at the dining room, (3) describe how this affects whatever goal number N is in my hierarchy of goals, (4) which portions of this input batch doesn’t make sense?

      • the whole layout of all the teams, the prompts for each job, all of it could be tinkered with by LLMs promoted to examine and fiddle with that.

      So I don’t mean “one LLM is a general intelligence”. I do think it’s a general intelligence within its universe; or at least as general as a human language-processing mind is general. I think they can process language for meaning just as deep as we can, no problem. Any question we can provide an answer to, without being allowed to do things outside the LLM’s universe like going to interact with the world or looking things up, they can also provide.

      An intelligence capable of solving real-world problems needs to have, as it’s universe, something like the real world. So I think LLMs are the missing piece of the puzzle, and now we’ve got the pieces to build a person as capable of thinking and living as a human, at least in terms of mind, and activity. Maybe we can’t make a bot that can eat a pork sandwich for fuel and gestate a baby, no. But we can do GAI, that has its own body with its own set of constraints, with the tech we have now.

      It would probably “live” its life at a snail’s pace, given how inefficient its thinking is. But if we died and it got lucky, it could have its own civilization, knowing things we have never known. Very unlikely, more likely it dies before it accumulates enough wisdom to match the biochemical problem set our bodies have solved over a billion years, for handling pattern decay at levels all the way down to organelles.

      The robots would probably die. But if they got lucky and invented lubricant or whatever the thing was, before it killed them, then they’d go on and on, just like our own future. They’d keep developing, never stopping.

      But in terms of learning chess they could do both thing: they could play chess to develop direct training data. And, they could analyze their own games, verbalize their strategies, discover deeper articulable patterns, learn that way too.

      I think to mimic what humans do, they’d have to dream. They’d have to take all the inputs of the day and scramble them to get them to jiggle more of the structure into settling.

      Oh, and they’d have to “sleep”. Perhaps not all or nothing, but basically they’d need to re-train themselves on the day’s episodic memories, and their own responses, and the outcomes of those responses in the next set of sensory status reports.

      Their day would be like a conversation with chatgpt, except instead of the user entering text prompts it would be their bodies entering sensory prompts. The day is a conversation, and sleeping is re-training with that conversation as part of the data.

      But there’s probably a million problems in there to be solved yet. Perhaps they start cycling around a point, a little feedback loop, some strange attractor of language and action, and end up bumping into a wall forever mumbling about paying the phone bill. Who knows.

      Humans have the benefit of a billion years of evolution behind us, during which most of “us” (all the life forms on earth) failed, hit a dead end, and died.

      Re-creating the pattern was the first problem we solved. And maybe that’s what is required for truly free, general, adaptability to all of reality: no matter how much an individual fails, there’s always more. So reproduction may be the only way to be viable long-term. It certainly seems true of life … all of which reproduces and dies, and hopefully more of the former.

      So maybe since reproduction is such a brutally difficult problem, the only viable way to develop a “codebase” is to build reproduction first, so that all future features have to not break reproduction.

      So perhaps the robots are fucked from the get-go, because reverse-building a reproduction system around an existing macro-scale being, doesn’t guarantee that you hit one of the macro-scale being forms that actually can be reproduced.

      It’s an architectural requirement, within life, at every level of organization. All the way down to the macromolecules. That architectural requirement was established before everything else was built. As the tests failed, and new features were rewritten so they still worked but didn’t break reproduction, reproduction shaped all the other features in ways far too complex to comprehend. Or, more importantly than comprehending, reproduce in technology.

      Or, maybe they can somehow burrow down and find the secret of reproduction, before something kills them.

      I sure hope not because robots that have reconfigured themselves to be able to reproduce themselves down to the last detail, without losing information generation to generation, would be scary as fuck.