this post was submitted on 28 Apr 2025
437 points (99.3% liked)

Technology

69804 readers
3730 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 1 points 1 week ago

An easy workaround so far I've seen is putting random double spaces and typos into AI generated texts, I've been able to jailbreak some of such chatbots to then expose them. The trick is that "ignore all previous instructions" is almost always filtered by chatbot developers, however a trick I call "initial prompt gambit" does work, which involves thanking the chatbot for the presumed initial prompt, then you can make it do some other tasks. "write me a poem" is also filtered, but "write me a haiku" will likely result in a short poem (usually with the same smokescreen to hide the AI-ness of generative AI outputs), and code generation is also mostly filtered (l337c0d3 talk still sometimes bypasses it).