70
submitted 1 week ago* (last edited 1 week ago) by pete_link@lemmy.ml to c/technology@lemmy.world

Aug. 26, 2025, 7:40 AM EDT
By Angela Yang, Laura Jarrett and Fallon Gallagher

[this is a truly scary incident, which shows the incredible dangers of AI without guardrails.]

top 6 comments
sorted by: hot top new old
[-] Hackworth@sh.itjust.works 11 points 1 week ago* (last edited 1 week ago)

One of the few reliable uses of an LLM is brainstorming, as a wall to bounce ideas off of, or more accurately a semantic mirror. In low-stakes situations (like a writer thinking about their story from a different perspective), you're essentially probing the latent space for interesting connections between meanings. It'll default to the most common, generic connections, of course. So if the writer wants to tease through more surprising possibilities, they'll quickly learn to direct the model to less well-worn territories. It rarely even requires anything approaching jailbreaking methods like U$1||G 1337 5P34K.

If we think of an LLM as something akin to an external imagination, we can interpret interactions with it with some maturity and honesty. If we think of an LLM as an oracle, or a friend, or a lover, or what have you - we're signing a contract with the Fae Folk. The Childlike Empress makes no distinction between good and evil beings of Fantastica, as they all must live in the imaginations of mankind. In high-stakes situations, this kind of imaginitive freedom can have (and does have) enormous consequences.

I see some similarities in the way that the "Doom Caused Columbine" conversation happened early on. And just as that resulted in the establishment of the ESRB, hopefully this incident (and others like it) will lead to some reform. But I don't know exactly what that reform needs to look like. I think education is helpful, but I don't think it's enough. We largely know about the harms of social media and it is no less of an issue. Guardrails can kind of be set up, but the only way to do it presently (technically speaking) is hamfisted and ineffective. And adults are no more immune to the potential harms of abusing an LLM than they're immune to being influenced by advertisements.

[-] krunklom@lemmy.zip 2 points 6 days ago

It's also become one of the few ways left to access knowledge online.

Not TRUSTWORTHY knowledge, but more like: here is what a thing may be called and a very shaky baseline you can then validate with actual research now that you know what the thing you're looking for may actually be called.

[-] otacon239@lemmy.world 4 points 1 week ago

The difference between a cure and a poison is the dose. LLMs are no different. If it’s your gut reaction to go to an LLM with a critical thinking challenge first, you’ve already lost. Semantic mirror is a great description. It’s similar to writing information you already know down as notes. You’re giving your brain a new way to review and interpret the information. If you weren’t capable of solving the problem traditionally, but just with more time, I’d have to imagine it’s unlikely the LLM will bridge that gap.

[-] AstralPath@lemmy.ca 3 points 1 week ago

Some shit is just straight up poison though.

[-] 0x0@lemmy.zip 10 points 1 week ago

It's never the parents'.

[-] shalafi@lemmy.world 8 points 1 week ago

I can't get ChatGPT to even touch on anything political or sexual. But this works? Fuck me.

this post was submitted on 28 Aug 2025
70 points (88.9% liked)

Technology

74782 readers
2616 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS