• Ace! _SL/S@ani.social
    link
    fedilink
    English
    arrow-up
    67
    arrow-down
    12
    ·
    edit-2
    11 months ago

    This comment sums it up pretty nicely:

    LOL innovative invention of swapping memory to storage…… maybe they can call it something cool like “cache”.

    Apple being “innovative” my ass, lmao

    • lemmylommy@lemmy.world
      link
      fedilink
      English
      arrow-up
      14
      arrow-down
      2
      ·
      11 months ago

      Well, if that commenter had more than just a vague idea of caching and/or swapping, they would know that the right algorithm can make or break performance.

      That paper is not “we invented caching”, but “this is how we make some certain models work well despite constraints imposed by RAM and flash storage.”

      It’s a worthy job for an engineer or researcher. Not quite as innovative as the invention of the wheel, but still enough to write a paper on (and read it, if you can manage to understand it).

    • Hegar@kbin.social
      link
      fedilink
      arrow-up
      13
      arrow-down
      3
      ·
      11 months ago

      The easiest way to tell that something’s not really innovative is if the person describing it uses the word innovative.

    • 4am@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      4
      ·
      11 months ago

      35 upvotes in the technology community…man you guys really are just all knee-jerk reactionaries and it really knowledgeable tech at all. git gud

  • guitarsarereal@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    26
    ·
    edit-2
    11 months ago

    Everyone likes to trash machine learning because the power requirements are high, but what they don’t realize is that we’re in the very first days of this technology (well, first couple decades of the technology being around, first few years of it being advanced enough to have anything to show off). Every technology that got bundled together into your phone was equally as useless when it was first invented. Honestly, compared to the development of most other technologies I’ve looked at, the pace of development in AI has been shocking.

    Literally once a week, I see some news story about AI researchers delivering an order of magnitude speedup in some aspect of AI inference. The technique described here apparently allows for a 20x speedup on GPU’s.

    • cybersandwich@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      ·
      11 months ago

      Whispercpp works off the ML cores on the m series chips. It’s faster than my 1080ti that I have in a server doing the same things–by orders of magnitude. And it sips power.

      Purpose built chips can be super powerful for their specific purposes.

  • fruitycoder@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    3
    ·
    11 months ago

    Tbh I’m more excited to see someone do use webnn, webgpu and petals together. Building smaller tighter models is good too.

  • wizzor
    link
    fedilink
    English
    arrow-up
    1
    ·
    11 months ago

    I don’t understand the innovation, I already run LLMs and stable diffusion on a laptop from 2011.

    I have no doubt it could be run on my Android phone.

  • Lophostemon@aussie.zone
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    29
    ·
    11 months ago

    Why the hell do we want to encourage people running MLMS on our phones?!! I don’t want to be part of some stupid pyramid scheme nonsense.