this post was submitted on 25 Jul 2024
1142 points (98.4% liked)
memes
10285 readers
4286 users here now
Community rules
1. Be civil
No trolling, bigotry or other insulting / annoying behaviour
2. No politics
This is non-politics community. For political memes please go to !politicalmemes@lemmy.world
3. No recent reposts
Check for reposts when posting a meme, you can only repost after 1 month
4. No bots
No bots without the express approval of the mods or the admins
5. No Spam/Ads
No advertisements or spam. This is an instance rule and the only way to live.
Sister communities
- !tenforward@lemmy.world : Star Trek memes, chat and shitposts
- !lemmyshitpost@lemmy.world : Lemmy Shitposts, anything and everything goes.
- !linuxmemes@lemmy.world : Linux themed memes
- !comicstrips@lemmy.world : for those who love comic stories.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Keep in mind that LLMs are essentially just large text predictors. Prompts aren't so much instructions as they are setting up the initial context of what the LLM is trying to predict. It's an algorithm wrapped around a giant statistical model where the statistical model is doing most of the work. If that statistical model is relied on to also control or limit the output of itself, then that control could be influenced by other inputs to the model.
Also they absolutely want the LLM to read user input and respond to it. Telling it exactly which inputs it shouldn't respond to is tricky.
In traditional programs this is done by "sanitizing input", which is done by removing the special characters and very specific keywords that are generally used when computers interpret that input. But in the case of LLMs, removing special characters and reserved words doesn't do much.