this post was submitted on 24 May 2025
126 points (87.1% liked)
Technology
70283 readers
3968 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I suspect most of the major models are as well. Kind of like how the Chinese models deal with Tienanmen Square.
Actually the Chinese models aren't trained to avoid Tiananmen Square. If you grabbed the model and ran it on your own machine, it will happily tell you the truth.
They censored their AI at a layer above the actual LLM, so users of their chat app would find results being censored.
That's...silly
Yes, they are. I only run LLMs locally and Deepseek R1 won't talk about Tiannamen square unless you trick it. They just implemented the protection badly.
Which would make sense from a censorship point of view as jailbreaks would be a problem. Just a filter/check before the result is returned for
*tiananmen*
is a much harder to break thing than guaranteeing the LLM doesn't get jailbroken/hallucinate.Wow... I don't use AI much so I didn't believe you.
The last time I got this response was when I got into a debate with AI about it being morally acceptable to eat dolphins because they are capable of rape...