Oh no
Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
Anyway
Oh no. Anyways...
Having now flooded the internet with bad AI content not surprisingly its now eating itself. Numerous projects that aren't AI are suffering too as the quality of text reduces.
is it not relatively trivial to pre-vet content before they train it? at least with aigen text it should be.
It's like a human centipede where only the first person is a human and everyone else is an AI. It's all shit, but it gets a bit worse every step.
Deep fired AI art sucks and is a decade late to the party
Good riddance.
I was very interested in the thumbnail of this post so I did a little digging and found this: The PDF to the Paper where the whole picture is
Wow, it's amazing that just 3.3% of the training set coming from the same model can already start to mess it up.
"Model collapse" is just a fancy way of saying "our stupid ideas are bad and nobody wants them."
Usually we get an AI winter, until somebody develops a model that can overcome that limitation of needing more and more data. In this case by having some basic understanding instead of just having a regurgitation engine for example. Of course that model runs into the limit of only having basic understanding, not advanced understanding and again there is an AI winter.
Sooner or later it is supposed to happen, but I don't think we are quite there....Yet.
Our wetware neutral networks probably aren't supposed to engage with synthetic content like this either. In a few years we're gonna learn that overexposure to AI generated content creates some sort of neurological problem in people, like a real-world "nerve attenuation syndrome" (Johnny Mnemonic).
Good
If we can work out which data conduits are patrolled more often by AI than by humans, we could intentionally flood those channels with AI content, and push Model Collapse along further. Get AI authors to not only vet for "true human content", but also pay licensing fees for the use of that content. And then, hopefully, give the fuck up on their whole endeavor.