this post was submitted on 26 Aug 2023
794 points (91.4% liked)

Programmer Humor

19197 readers
978 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 1 year ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] NounsAndWords@lemmy.world 25 points 1 year ago (3 children)

It just occurred to me that AI in the nearish future will probably/almost certainly be able to do this.

[–] Psythik@lemm.ee 36 points 1 year ago (1 children)

I can't wait for AI to make a PC port of every console game ever so that we can finally stop using emulators.

[–] amki@feddit.de 8 points 1 year ago (6 children)

This won't happen in our lifetime. Not only because this is more complex than rambling vaguely correlated human speech while hallucinating half the time.

[–] secret301@sh.itjust.works 4 points 1 year ago

I think it'll be in our lifetime just not anytime soon. I feel like AI is gonna boom like the internet did. Didn't happen overnight and not even in a year but over 35ish years

[–] SnipingNinja@slrpnk.net 3 points 1 year ago (1 children)

Idk the specifics, but what you say makes it sound like it would be easier to create an AI that recreates a game based on gameplay visuals (and the relevant controls)

[–] amki@feddit.de 1 points 1 year ago

That game would still not work because there is a ton of hidden state in all but the simplest computer games that you cannot tell from just playing through the game normally.

An AI could probably reinvent flappy birds because there is no more depth than what is currently on screen but that's about it.

[–] Mockrenocks@lemmy.world 2 points 1 year ago

Ai prompt: make me a program that will convert PS5 games to PC

AI: Use Convert-PS5GameToPC

End of line

[–] i_am_hiding@aussie.zone 0 points 1 year ago (1 children)

AI can literally read minds. I don't think it's that great of a step to say it should be able to decompile a few games.

[–] amki@feddit.de 1 points 1 year ago (1 children)

About half the time, the text closely – and sometimes precisely – matched the intended meanings of the original words.

Don't be surprised but about half of the time I can predict the result of a coin flip.

I'm not saying it's not interesting but needing custom training and an fMRI is not "an AI can read minds"

It can see if patterns it saw previously reappear in a heavily time delayed fMRI. Looking for patterns you already know isn't such an impressive feat Computers have done this for ages now.

It litterally can't read minds.

[–] sfgifz@lemmy.dbzer0.com 2 points 1 year ago* (last edited 1 year ago) (1 children)

Later, the same participants were scanned listening to a new story or imagining telling a story and the decoder was used to generate text from brain activity alone. About half the time, the text closely – and sometimes precisely – matched the intended meanings of the original words.

You left out the most important context about "half of the time". Guessing what you're thinking of by just looking at your brain activity with a 50% accuracy is a very very good achievement - it's not pulling it out of a 1 or 0 outcome like you're with your coin flip.

You can pretend that the AI is useless and you're the smartest boy in the class all you want, doesn't negate the accomplishments.

[–] amki@feddit.de 1 points 1 year ago

Being close (and "sometimes" precise) to the intended meaning is an equally useless metric to measure performance.

Depending on what you allow for "well close enough I think" asking ChatGPT to tell a story without any reading of fMRI would get you to these results. Especially if you know beforehand it's gonna be a story told.

[–] perviouslyiner@lemm.ee 9 points 1 year ago* (last edited 1 year ago) (2 children)

It was a staple of Asimov's books that while trying to predict decisions of the robot brain, nobody in that world ever understood how they fundamentally worked.

He said that while the first few generations were programmed by humans, everything since that was programmed by the previous generation of programs.

This leads us to Asimov's world in which nobody is even remotely capable of creating programs that violate the assumptions built into the first iteration of these systems - are we at that point now?

[–] amki@feddit.de 8 points 1 year ago (2 children)

No. Programs cannot reprogram themselves in a useful way and are very very far from it.

[–] legion02@lemmy.world 1 points 1 year ago

Eh, I'd say continuous training models are pretty close to this. Adapting to changing conditions and new input is kinda what they're for.

[–] yum13241@lemm.ee 1 points 1 year ago (1 children)

Then how does polymorphic/self-modifying code work?

[–] amki@feddit.de 1 points 1 year ago

It doesn't or do you have serious applications for self-modifying code?

[–] rikudou@lemmings.world 2 points 1 year ago

are we at that point now?

Nope, but we're getting there.

[–] amki@feddit.de 1 points 1 year ago