First off, the beauty of these two posts being beside each other is palpable.
Second, as you can see on the picture, it's more like 60%
People tweeting stuff. We allow tweets from anyone.
RULES:
First off, the beauty of these two posts being beside each other is palpable.
Second, as you can see on the picture, it's more like 60%
No it's not. If you actually read the study, it's about AI search engines correctly finding and citing the source of a given quote, not general correctness, and not just the plain model
Read the study? Why would i do that when there's an infographic right there?
(thank you for the clarification, i actually appreciate it)
I use chatgpt as a suggestion. Like an aid to whatever it is that I’m doing. It either helps me or it doesn’t, but I always have my critical thinking hat on.
I did a google search to find out how much i pay for water, the water department where I live bills by the MCF (1,000 cubic feet). The AI Overview told me an MCF was one million cubic feet. It's a unit of measurement. It's not subjective, not an opinion and AI still got it wrong.
Everywhere else in the world a big M means million.
I think in this case it's Roman numeral M
The only thing that would make more sense would be if the bill was in cuneiform.
40% seems low
that depends on what topic you know and how well you know it.
LLMs are actually pretty good for looking up words by their definition. But that is just about the only topic I can think of where they are correct even close to 80% of the time.
Most of my searches have to do with video games, and I have yet to see any of those AI generated answers be accurate. But I mean, when the source of the AI's info is coming from a Fandom wiki, it was already wading in shit before it ever generated a response.
I’ve tried it a few times with Dwarf Fortress, and it was always horribly wrong hallucinated instructions on how to do something.
I've been using o3-mini mostly for ffmpeg
command lines. And a bit of sed
. And it hasn't been terrible, it's a good way to learn stuff I can't decipher from the man pages. Not sure what else it's good for tbh, but at least I can test and understand what it's doing before running the code.
I just use it to write emails, so I declare the facts to the LLM and tell it to write an email based on that and the context of the email. Works pretty well but doesn't really sound like something I wrote, it adds too much emotion.
This is what LLMs should be used for. People treat them like search engines and encyclopedias, which they definitely aren't
That sounds like more work than just writing the email to me
Yeah, that has been my experience so far. LLMs take as much or more work vs the way I normally do things.
If you want an AI to be an expert, you should only feed it data from experts. But these are trained on so much more. So much garbage.
This, but for tech bros.