One of the issues with LLM is that it attracted all attention. Classifiers are generally cool, cheap and saved us from multiple issues (ok face recognition aside 🙂)
When the AI bubble will burst (because of LLM being expensive and not good enough to replace a person even if they are good in pretending to be a person) all AI will slow down… including classifiers, NLP, etc
All this because the AI community was obsessed by the Turing test/imitation game 🙄
Turing was a genius but heck if I am upset with him for coming with this BS 🤣
It made sense in the context it was devised in. Back then we thought the way to build an AI was to build something that was capable of reasoning about the world.
The notion that there’d be this massive amount of text generated by a significant percentage of the world’s population all typing their thoughts into networked computers for a few decades, coupled with the digitisation of every book written, that could be stitched together in a 1,000,000,000,000-byte model that just spat out the word with the highest chance of being next based on what everyone else in the past had written, producing the illusion of intelligence, would have been very difficult for him to predict.
Remember, Moore’s Law wasn’t coined for another 15 years, and personal computers didn’t even exist as a sci-fi concept until later still.
One of the issues with LLM is that it attracted all attention. Classifiers are generally cool, cheap and saved us from multiple issues (ok face recognition aside 🙂)
When the AI bubble will burst (because of LLM being expensive and not good enough to replace a person even if they are good in pretending to be a person) all AI will slow down… including classifiers, NLP, etc
All this because the AI community was obsessed by the Turing test/imitation game 🙄
Turing was a genius but heck if I am upset with him for coming with this BS 🤣
It made sense in the context it was devised in. Back then we thought the way to build an AI was to build something that was capable of reasoning about the world.
The notion that there’d be this massive amount of text generated by a significant percentage of the world’s population all typing their thoughts into networked computers for a few decades, coupled with the digitisation of every book written, that could be stitched together in a 1,000,000,000,000-byte model that just spat out the word with the highest chance of being next based on what everyone else in the past had written, producing the illusion of intelligence, would have been very difficult for him to predict.
Remember, Moore’s Law wasn’t coined for another 15 years, and personal computers didn’t even exist as a sci-fi concept until later still.
I dunno about that. We got a pile of architecture research out of it just waiting for some more tests/implementations.
And think of how cheap renting compute will be! It’s already basically subsidized, but imagine when all those A100s/H100s are dumped.