this post was submitted on 21 Jun 2025
265 points (95.2% liked)

AI

5099 readers
17 users here now

Artificial intelligence (AI) is intelligence demonstrated by machines, unlike the natural intelligence displayed by humans and animals, which involves consciousness and emotionality. The distinction between the former and the latter categories is often revealed by the acronym chosen.

founded 4 years ago
 

Tony Stark speedrunning model collapse.

you are viewing a single comment's thread
view the rest of the comments
[–] NocturnalMorning@lemmy.world 24 points 2 days ago (3 children)

Isn't it a well known fact that training on other AI output data leads to complete collapse of the newly trained AI models?

[–] LodeMike@lemmy.today 13 points 2 days ago (1 children)

Do you think this guy knows facts?

[–] NocturnalMorning@lemmy.world -5 points 2 days ago (3 children)

Yes, sometimes. He has a physics degree, he's not an idiot exactly. Just extremely vain and narcissistic.

[–] LodeMike@lemmy.today 1 points 2 days ago (1 children)

Why are you being downvoted :(

I was just trying to pull your leg.

[–] NocturnalMorning@lemmy.world 2 points 2 days ago

🤷‍♀️ Dunno, people are weird sometimes.

[–] LodeMike@lemmy.today 1 points 2 days ago* (last edited 2 days ago)

What was the major? Judging by what he does it'd probably be "kinematics of sex"

[–] iAvicenna@lemmy.world 1 points 2 days ago

From what he wrote it feels like it will majorly be existing data with substitutions/corrections made in places where they deem necessary. Like when you ask about Elon it will probably spew sth along the lines of the greatest inventor of the last century, a polymath and a very successful path of exile 2 player.

[–] 8uurg@lemmy.world 1 points 2 days ago

Not quite, actually. It is moreso training recursively on the output without any changes, i.e., Data -> Model A -> Data (generated by Model A) -> Model B -> Data (generated by Model B -> ..., that leads to (complete) collapse. A single step like this can still worsen performance notably, though, especially when it makes up the sheer majority of the data. [source]

And if they train using little data, you won't get anywhere near the chatbots we have now. If they fine-tune an existing model to do as they wish, it would likely have side effects. Like being more likely to introduce security bugs in generated code, generally give incorrect answers to other common sense questions, and so on. [source]