this post was submitted on 12 Jun 2024
114 points (98.3% liked)
Quark's
1097 readers
1 users here now
Come to Quark’s, Quark’s is Fun!
General off-topic chat for the crew of startrek.website. Trek-adjacent discussions, other sci-fi television, navigating the Fediverse, server meta (within reason), selling expired cases of Yamok sauce, it’s all fair game.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
What will it take until people get it through their thick skulls that ChatGPT isn't intelligent, doesn't learn and is a tool that can only generate plausible gibberish.
Using the same tools to detect such gibberish will give you more gibberish.
Garbage in, Garbage out has been true since the difference engine, it's just that today the garbage smells like English words, still garbage, but not knowledge, intelligence or anything like it.
The machine learning approach for building models, used to produce so called large language models like ChatGPT is also used to create weather forecasting models that are bigger, better and orders of magnitude faster than available until now.
The tools have changed life, but I'm unconvinced that it's a suitable, sustainable or realistic way to create artificial intelligence, despite claims to the contrary.
People are so insistent that it's ai that it all reminds me of Blockchain. It's new! It'll change everything!
It'll change some things. What we are seeing now is business forcing it into everything when really, right now, there are only a handful of things it makes sense to use.
It's really great at giving you a starting point a very rough outline of something. That is the easy part. The hard part is turning that into something new and coherent, and for that I think modern AI is nowhere close. That needs a human
It is by definition AI
In the sense that AI is an extremely general term that involves many different technologies, yes. Generative AI/LLMs are not true AGI, which is what people think it is. It cannot think, it cannot learn, it can only predict.
People think it AI intelligence is comparable to how a hovercraft hovers, as in the word is taken literally, but it is actually comparable to a Hoverboard.
That's actually pretty good... the techbro equivalent of "We did it!"
That's a distinction without a difference. If it can predict what a AGI would do in a given situation, then it is an AGI.
I'm not saying that it is an AGI, but the reason it it isn't is more than "it can only predict".