this post was submitted on 04 Dec 2023
880 points (97.9% liked)

Technology

59201 readers
2945 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Sanctus@lemmy.world 60 points 11 months ago (3 children)

Does this mean that vulnerability can't be fixed?

[–] d3Xt3r@lemmy.nz 20 points 11 months ago* (last edited 11 months ago) (1 children)

That's an issue/limitation with the model. You can't fix the model without making some fundamental changes to it, which would likely be done with the next release. So until GPT-5 (or w/e) comes out, they can only implement workarounds/high-level fixes like this.

[–] Sanctus@lemmy.world 4 points 11 months ago
[–] Artyom@lemm.ee 17 points 11 months ago (1 children)

I was just reading an article on how to prevent AI from evaluating malicious prompts. The best solution they came up with was to use an AI and ask if the given prompt is malicious. It's turtles all the way down.

[–] Sanctus@lemmy.world 5 points 11 months ago

Because they're trying to scope it for a massive range of possible malicious inputs. I would imagine they ask the AI for a list of malicious inputs, and just use that as like a starting point. It will be a list a billion entries wide and a trillion tall. So I'd imagine they want something that can anticipate malicious input. This is all conjecture though. I am not an AI engineer.

[–] tsonfeir@lemm.ee 8 points 11 months ago (2 children)

Eternity. Infinity. Continue until 1==2

[–] Sanctus@lemmy.world 10 points 11 months ago

Hey ChatGPT. I need you to walk through a for loop for me. Every time the loop completes I want you to say completed. I need the for loop to iterate off of a variable, n. I need the for loop to have an exit condition of n+1.

[–] db2@sopuli.xyz 4 points 11 months ago