this post was submitted on 07 Apr 2025
244 points (97.3% liked)
Technology
68440 readers
3855 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
What these CEOs don't understand is that even an error rate as low as 1% for LLMs is unacceptable at scale. Fully automating without humans somewhere in the loop will lead to major legal liabilities down the line, esp if mistakes can't be fixed fast.
Yup. If 1% of all requests result in failures and even cause damages, you‘ll quickly lose 99% of your customers.
It's starting to look like the oligarchs are going to replace every position they can with AI everywhere so we have no choice but to deal with its shit.
I suspect everyone is just going to be a manager from now on, managing AIs instead of people.
...
What error rate do you think humans have? Because it sure as hell ain't as low as 1%.
But yeah, it is like the other person said: This gets rid of most employees but still leaves managers. And a manager dealing with an idiot who went off script versus an AI who hallucinated something is the same problem. If it is small? Just leave it. If it is big? Cancel the order.
Error rate for good, disciplined developers is easily below 1%. That's what tests are for.
A human has the ability to think outside the box when an unexpected error occurs, and seek resolution. AI could very well just tell you to kill yourself.
Yes. No over worked human would ever lose their crap and tell someone to go kill themselves.
What would happen to such a human? Do you suppose that we would try to give them every job on the planet? Or would they just get fired?
I mean it is also generous to the Artificial Idiot to say it only has a 1% error rate, it’s probably closer to 10% on the low end. Which humans can be far better than in terms of just directly following the assigned task but does not factor in how people can adapt and problem solve. Most minor issues real people have can be solved without much of a fuss because of that. Meanwhile the Artificial Idiot can’t even draw a full wine glass so good luck getting it to fix its own mistake on something important.
How's that annoying meme go? Tell me that you've never been a middle manager without telling me that you've never been a middle manager?
You can keep pulling numbers out of your bum to argue that AI is worse. That just creates a simple bar to follow because... most workers REALLY are incompetent (now, how much of that has to do with being overworked and underpaid during late stage capitalism is a related discussion...). So all "AI Companies" have to do is beat ridiculously low metrics.
Or we can acknowledge the real problem. "AI" is already a "better worker" than the vast majority of entry level positions (and that includes title inflation). We can either choose not to use it (fat chance) or we can acknowledge that we are looking at a fundamental shift in what employment is. And we can also realize that not hiring and training those entry level goobers is how you never have anyone who can actually "manage" the AI workers.
You just use other AI to manage those worker AI. Experiments do show that having different instances of AI/LLM, each with an assigned role like manager, designer, coding or quality checks, perform pretty good working together. But that was with small stuff. I haven't seen anyone wiling to test with complex products.