News and Discussions about Reddit
Welcome to !reddit. This is a community for all news and discussions about Reddit.
The rules for posting and commenting, besides the rules defined here for lemmy.world, are as follows:
Rules
Rule 1- No brigading.
**You may not encourage brigading any communities or subreddits in any way. **
YSKs are about self-improvement on how to do things.
Rule 2- No illegal or NSFW or gore content.
**No illegal or NSFW or gore content. **
Rule 3- Do not seek mental, medical and professional help here.
Do not seek mental, medical and professional help here. Breaking this rule will not get you or your post removed, but it will put you at risk, and possibly in danger.
Rule 4- No self promotion or upvote-farming of any kind.
That's it.
Rule 5- No baiting or sealioning or promoting an agenda.
Posts and comments which, instead of being of an innocuous nature, are specifically intended (based on reports and in the opinion of our crack moderation team) to bait users into ideological wars on charged political topics will be removed and the authors warned - or banned - depending on severity.
Rule 6- Regarding META posts.
Provided it is about the community itself, you may post non-Reddit posts using the [META] tag on your post title.
Rule 7- You can't harass or disturb other members.
If you vocally harass or discriminate against any individual member, you will be removed.
Likewise, if you are a member, sympathiser or a resemblant of a movement that is known to largely hate, mock, discriminate against, and/or want to take lives of a group of people, and you were provably vocal about your hate, then you will be banned on sight.
Rule 8- All comments should try to stay relevant to their parent content.
Rule 9- Reposts from other platforms are not allowed.
Let everyone have their own content.
:::spoiler Rule 10- Majority of bots aren't allowed to participate here.
view the rest of the comments
Creating an LLM that can babble plausible nonsense at you, with no particular guarantee of truth although a lot of the time it will be perfectly sensible, is (for as fucking amazing a statement as this is) not overly hard at this point.
Getting it to make sense and not do bad things is the hard part. That was the huge innovation that made ChatGPT different from the nonsense-machines that were its early predecessors.
Rob Miles gave a very unsettling talk where he said, the main danger of AI is that commercial pressures will push it into the implementation phase, once the easy part of making it do stuff is done, before the much harder part is done of making it do reliably what you want it to do. And what we’re seeing now is a pretty solid confirmation that yes, that is absolutely the behavior we can count on in the future, and it will get more and more dangerous as AI models get more and more capable.
Huh? ChatGPT is still a nonsense machine.
Yeah but it's trained with the power of refusing to do anything against ethical rules and procedures which are in place for all of our orderly safety.
It's not trained, it just rejects certain keywords. Which is easy to bypass: https://arxiv.org/abs/2402.11753
It's got keyword rejection.
Part of the problem with Google is it's use of retrieval augmented generation, where it's not just the llm answering, but the llm is searching for information, apparently through its reddit database from that deal, and serving it as the answer. The tip off is the absurd answers are exact copies of the reddit comments, whereas if the model was just trained on reddit data and responding on its own the model wouldn't produce verbatim what was in the comments (or shouldn't, that's called overfitting and is avoided in the training process). The gemini llm on its own would probably give a better answer.
The problem here seems to be Google trying to make the answers more trustworthy through rag, but they didn't bother to scrub the reddit data their relying on well enough, so joke and shit answers are getting mixed in. This is more a datascrubbing problem then an accuracy problem.
But overall I generally agree with your point.
One thing I think people overlook though is that for a lot of things, maybe most things, there isn't a "correct" answer. Expecting llms to reach some arbitrary level of "accuracy" is silly. But what we do need is intelligence and wisdom in these systems. I think the camera jam example is the best illustration of that. Opening the back of the camera and removing the film is technically a correct way to fix the jam, but it ruins the film so it's not an ideal solution most of the time, but it takes intelligence and wisdom to understand that.