this post was submitted on 12 Jul 2025
20 points (100.0% liked)

TechTakes

2077 readers
35 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
 

‘an expensive slot machine that outputs slop 98% of the time’

podcast and blog post tomorrow, i am ill

yes that's a clothes hanger at the back i forgot to put in the hall, we only reveal our clean laundry here

you are viewing a single comment's thread
view the rest of the comments
[–] HedyL@awful.systems 4 points 4 days ago* (last edited 4 days ago)

It's quite noteworthy how often these shots start out somewhat okay at the first prompt, but then deteriorate markedly over the following seconds.

As a layperson, I would try to explain this as follows: At the beginning, the AI is - to some extent - free to "pick" how the characters and their surroundings would look like (while staying within the constraints of the prompt, of course, even if this doesn't always work out either).

Therefore, the AI can basically "fill in the blanks" from its training data and create something that may look somewhat impressive at first glance.

However, for continuing the shot, the AI is now stuck with these characters and surroundings while having to follow a plot that may not be represented in its training data, especially not for the characters and surroundings it had picked. This is why we frequently see inconsistencies, deviations from the prompt or just plain nonsense.

If I am right about this assumption, it might be very difficult to improve these video generators, I guess (because an unrealistic amount of additional training data would be required).

Edit: According to other people, it may also be related to memory/hardware etc. In that case, my guesses above may not apply. Or maybe it is a mixture of both.