LLM is what usually sold as AI nowadays. Convential ML is boring and too normal, not as exciting as a thing that processes your words and gives some responses, almost as if it’s sentient. Nvidia couldn’t come to it’s current capitalization if we defaulted to useful models that can speed up technical process after some fine tuning by data scientists, like shaving off another 0.1% on Kaggle or IRL in a classification task. It usually causes big but still incremental changes. What is sold as AI and in what quality it fits into your original comment as a lifesaver is nothing short of reinvention of one’s workplace or completely replacing the worker. That’s hardly hapening anytime soon.
LLM is what usually sold as AI nowadays. Convential ML is boring and too normal, not as exciting as a thing that processes your words and gives some responses, almost as if it’s sentient.
To be fair, that’s because there are a lot of automation situations where having semantic understanding of a situation can be extremely helpful in guiding action over a ML model that is not semantically aware.
The reason that AI video generation and out painting is so good for instance it that it’s analyzing a picture and dividing it into human concepts using language and then using language to guide how those things can realistically move and change, and then applying actual image generation. Stuff like Waymo’s self driving systems aren’t being run through LLMs but they are machine learning models operating on extremely similar principles to build a semantic understanding of the driving world.
I’d argue, that it sometimes adds complexity to an already fragile system. Like when we implement touchscreens instead of buttons in cars. It’s akin to how Tesla, unlike Waymo, dropped LIDAR to depend on regular videoinputs alone. Direct control over systems without unreliable interfaces, semantic translation layer, computer vision dependancy etc serves the same tasks without additional risks and computational overheads.
LLM is what usually sold as AI nowadays. Convential ML is boring and too normal, not as exciting as a thing that processes your words and gives some responses, almost as if it’s sentient. Nvidia couldn’t come to it’s current capitalization if we defaulted to useful models that can speed up technical process after some fine tuning by data scientists, like shaving off another 0.1% on Kaggle or IRL in a classification task. It usually causes big but still incremental changes. What is sold as AI and in what quality it fits into your original comment as a lifesaver is nothing short of reinvention of one’s workplace or completely replacing the worker. That’s hardly hapening anytime soon.
To be fair, that’s because there are a lot of automation situations where having semantic understanding of a situation can be extremely helpful in guiding action over a ML model that is not semantically aware.
The reason that AI video generation and out painting is so good for instance it that it’s analyzing a picture and dividing it into human concepts using language and then using language to guide how those things can realistically move and change, and then applying actual image generation. Stuff like Waymo’s self driving systems aren’t being run through LLMs but they are machine learning models operating on extremely similar principles to build a semantic understanding of the driving world.
I’d argue, that it sometimes adds complexity to an already fragile system. Like when we implement touchscreens instead of buttons in cars. It’s akin to how Tesla, unlike Waymo, dropped LIDAR to depend on regular videoinputs alone. Direct control over systems without unreliable interfaces, semantic translation layer, computer vision dependancy etc serves the same tasks without additional risks and computational overheads.