Technology
This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.
Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.
Rules:
1: All Lemmy rules apply
2: Do not post low effort posts
3: NEVER post naziped*gore stuff
4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.
5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)
6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist
7: crypto related posts, unless essential, are disallowed
view the rest of the comments
The article makes several claims and insinuations without backing them up so I find it hard to follow any of the reasoning.
I don't think it's desirable that it's easier to reason about an AI than about a human. If it is, then we haven't achieved human-level intelligence. I posit that human intelligence can be reasoned about given enough understanding but we're not there yet, and until we are we shouldn't expect to be able to reason about AI either. If we could, it's just a sign that the AI is not advanced enough to fulfill its purpose.
Postel's law IMHO is a big mistake - it's what gave us Internet Explorer and arbitrary unpredictable interpretation of HTML, leading to decades of browser incompatibility problems. But the law is not even applicable here. Unlike the Internet, we want the AI to appear to think for itself rather than being predictable.
"Today's highly-hyped generative AI systems (most famously OpenAI) are designed to generate bullshit by design." Uh no? They're designed with the goal to generate useful content. The bullshit is just an unfortunate side effect because today's AI algorithms have not evolved very far yet.
If I had to summarize this article in one word, that would be it: bullshit.
I agree that the author didn't do a great job explaining, but they are right about a few things.
Primarily, LLMs are not truth machines. That just flatly and plainly not what they are. No researcher, not even OpenAI makes such a claim.
The problem is the public perception that they are. Or that they almost are. Because a lot of time, they're right. They might even be right more frequently than some people's dumber friends. And even when they're wrong, they sound right. Even when it's wrong, it still sounds smarter than most peoples smartest friends.
So, I think that the point is that there is a perception gap between what LLMs are, and what people THINK that they are.
As long as the perception is more optimistic than the reality, a bubble of some kind will exist. But just because there is a "reckoning" somewhere in the future doesn't imply it will crash to nothing. It just means the investment will align more closely to realistic expectations as the clarity of what realistic expectations even are become more clear.
LLMs are going to revolutionize and also destroy many industries. It will absolutely fundamentally change the way we interact with technology. No doubt...but for applications which strictly demand correctness, they are not appropriate tools. And investors don't really understand that yet.
OpenAI’s algorithm like all LLM’s is designed to give you the next most likely word in a sentence based on what most frequently came next in its training data. Their main strategy has actually been to use a older and simpler transformer algorithm, and to just vastly increase the scrapped text content and recently bias with each new release.
I would argue that any system that works by stringing sudorandom words together based on how often they appear in its input sources is not going to be able to do anything but generate bullshit, albeit bullshit that may happen to be correct by pure accident when it’s near directly quoting said input sources.
"Article". I'm going to call it what it is: a blog post that should have moderated away. If people here are going to post "tech news", make sure it has actual journalism.
It's almost like Isaac Asimov wrote a famous book about robotic laws and a bunch of different short stories on how easy it was to circumvent them.