There’s plenty of stuff where ML algorithms the state of the art. For example the raw data from nanopore DNA sequencing machines is extremely noisy and ML algorithms clean it up with much less error than the Markov chains used in years previous.
Coral*
Working with pretrained models implemented in FPGAs for particle identification and tracking. It’s much faster and exactly as accurate. ¯\_(ツ)_/¯
Run, the butlerian jihad is already going your way.
The actual model required for general purpose likely lies beyond the range of petabytes of memory.
These models are using gigabytes and the trend indicates its exponential. A couple more gigabytes isn’t going to cut it. Layers cannot expand the predictive capabilities without increasing the error. I’m sure a proof of that will be along within in the next few years.
“Come on man, I just need a couple more pets of your data and I will totally be able to predict you something useful!”.
It’s capacitors flip polarity in anticipation.“I swear man! It’s only a couple of orders of magnitude more, man! And all your dreams will come true. I’m sure I’ll service you right!”
Well if it needs it, right?
Ai sucks ass, stop using it
It doesn’t. It’s just overhyped.
“There is no free lunch.”, is a saying in ML research.
That’s just a saying.
Source?
For the meme? The Walking Dead. For the content? No idea.
So what you’re saying, Dad, is it’s nascent and already faster? Gotcha.