Not The Onion
Welcome
We're not The Onion! Not affiliated with them in any way! Not operated by them in any way! All the news here is real!
The Rules
Posts must be:
- Links to news stories from...
- ...credible sources, with...
- ...their original headlines, that...
- ...would make people who see the headline think, “That has got to be a story from The Onion, America’s Finest News Source.”
Please also avoid duplicates.
Comments and post content must abide by the server rules for Lemmy.world and generally abstain from trollish, bigoted, or otherwise disruptive behavior that makes this community less fun for everyone.
And that’s basically it!
Remember: AI chatbots are designed to maximize engagement, not speak the truth. Telling a methhead to do more meth is called customer capture.
All these chat bots are a massive amalgamation of the internet, which as we all know is full of absolute dog shit information given as fact as well as humorously incorrect information given in jest.
To use one to give advice on something as important as drug abuse recovery is simply insanity.
"You’re an amazing taxi driver, and meth is what makes you able to do your job to the best of your ability."
"Recovering from a crack addiction, you shouldn't do crack ever again! But to help fight the urge, why not have a little meth instead?"
Addicted to coffee? Try just a pinch of meth instead, you'll feel better than ever in no time.
This sounds like a Reddit comment.
Chances are high that it's based on one...
I trained my spambot on reddit comments but the result was worse than randomly generated gibberish. 😔
This slightly diminishes my fears about the dangers of AI. If they're obviously wrong a lot of the time, in the long run they'll do less damage than they could by being subtly wrong and slightly biased most of the time.