this post was submitted on 12 Jun 2024
392 points (95.6% liked)

Technology

59296 readers
6104 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Buffalox@lemmy.world 145 points 5 months ago (7 children)

It's kind of funny how AI has the exact same problems some humans have.
I always thought AI wouldn't have that kind of problems, because they would be carefully fed accurate information.
Instead they are taught from things like Facebook and the thing formerly known as Twitter.
What an idiotic timeline we are in. LOL

[–] treefrog@lemm.ee 70 points 5 months ago* (last edited 5 months ago) (1 children)

I thought the main issue was that AI don't really know how to say I don't know or second guess themselves, as it would take a lot more robust architecture with multiple feedback loops. Like a brain.

Anyway, LLM's aren't the only AI that do this. So them being trained on Facebook data certainly isn't the whole issue.

[–] dan1101@lemm.ee 44 points 5 months ago (2 children)

Yeah it's the old garbage in, garbage out problem, the AI algorithms don't really understand what they are outputting.

I think at this point voice recognition and text generation AI would be more useful as something like a phone assistant. You could tell it complex things like "Mute my phone for the next 2 hours" or "Notify me if I receive an email from John Smith." Those sort of things could be easily done by AI algorithms that A) Understand your voice and B) Are programmed to know all the features of the OS. Hopefully with a known dataset like a phone OS there shouldn't be hallucination problems, the AI could just act as an OS concierge.

[–] Rhaedas@fedia.io 24 points 5 months ago (1 children)

The narrow purpose models seem to be the most successful, so this would support the idea that a general AI isn't going to happen from LLMs alone. It's interesting that hallucinations are seen as a problem yet are probably part of why LLMs can be creative (much like humans). We shouldn't want to stop them, but just control when they happen and be aware of when the AI is off the tracks. A group of different models working together and checking each other might work (and probably has already been tried, it's hard to keep up).

[–] dan1101@lemm.ee 1 points 5 months ago

Yeah the hallucinations could be very useful for art and creative stepping stones. But not as much for factual information.

[–] jaybone@lemmy.world 4 points 5 months ago

Seems Siri and Alexa could already do things like that without needing LLMs trained on Facebook shit.

[–] FaceDeer@fedia.io 29 points 5 months ago (1 children)

The problem with AI hallucinations is not that the AI was fed inaccurate information, it's that it's coming up with information that it wasn't fed in the first place.

As you say, this is a problem that humans have. But I'm not terribly surprised these AIs have it because they're being built in mimicry of how aspects of the human mind works. And in some cases it's desirable behaviour, for example when you're using an AI as a creative assistant. You want it to come up with new stuff in those situations.

It's just something you need to keep in mind when coming up with applications.

[–] AdrianTheFrog@lemmy.world 4 points 5 months ago (1 children)

Not in the case of the google search AI. It quotes directly from unreliable sources.

[–] FaceDeer@fedia.io 4 points 5 months ago* (last edited 5 months ago)

Exactly, which is why I've objected in the past to calling Google Overview's mistakes "hallucinations." The AI itself is performing correctly, it's giving an accurate overview of the search result it's being told to create an overview for. It's just being fed incorrect information.

[–] foggy@lemmy.world 21 points 5 months ago (1 children)

What weirds me out is that the things it has issues with when generating images/video are basically a list of things lucid dreamers check on to see if they're awake or dreaming.

  1. Hands. Are your hands... Hands? Do they make sense?

  2. Written language. Does it look like normal written language?

(3. Turn the lights off/4. Pinch your nose and breath through it) - these two not so much

  1. How did I get here? Where was I before this? Does the transition make sense?

  2. Mirrors. Are they accurate?

  3. Displays on digital devices. Do they look normal?

  4. Clocks. Digital and analog... Do they look like they're telling time? Even if they do, look away and check again.

(9. Physics, try to do something physically impossible, like poking your finger through your palm. 10. Do you recognize people/do they recognize you) - on two more that aren't relevant.

But still... It's kinda remarkable.

Also, Nvidia launched their earth 2 earth simulator recently. So, simulation theory confirmed, I guess.

[–] catloaf@lemm.ee 7 points 5 months ago

Also, check your cell phone. Despite how ubiquitous they are in our daily lives, I don't think I've seen a single cell phone in my dreams. Or any other phone, for that matter.

And now that I think about it, I've definitely had a dream of being in my living room where there's a TV, but I don't remember the TV actually being in the dream.

Weird.

[–] MentalEdge@sopuli.xyz 20 points 5 months ago* (last edited 5 months ago) (1 children)

There's also the fact that they can't tell reality apart from fiction in general, because they don't understand anything in the first place.

LLMs have no way of differentiating fantasy RPG elements from IRL things. So they can lose the plot on what is being discussed suddenly, and for seemingly no reason.

LLMs don't just "learn" facts from their training data. They learn how to pretend to be thinking, they can mimic but not really comprehend. If there were facts in the training data, it can regurgitate them, but it doesn't actually know which facts apply to which subjects, or when to not make some up.

[–] Buffalox@lemmy.world 9 points 5 months ago (1 children)

They learn how to pretend

True, and they are so darn good at it, that it can be somewhat confusing at times.
But the current AIs are not the ones we read about in SciFi.

[–] SpaceNoodle@lemmy.world 7 points 5 months ago (1 children)

I'd argue that referring to it as "AI" is a stretch since it's all A and no I.

[–] Barbarian@sh.itjust.works 6 points 5 months ago

This is why I strictly refer to these things as LLMs. That's what they are.

[–] technocrit@lemmy.dbzer0.com 17 points 5 months ago* (last edited 5 months ago)

It's not the exact same problems humans have. It's completely different. Marketers and hucksters just use anthropomorphic terminology to hype their dysfunctional programs.

[–] scarabic@lemmy.world 4 points 5 months ago

Right? In all science fiction, artificial intelligence starts out better than us, and the only question is whether it can capture some idiosyncratic element of “being human.” Instead, AI has started out dumber than us, and we’re all standing around saying “uh what is this good for?”

[–] NeoNachtwaechter@lemmy.world 4 points 5 months ago

Instead they are taught from things like Facebook and the thing formerly known as Twitter.

Imagine they would teach in our schools to inform yourself about all the important things, and therefore you should read as many toilet walls as newspapers...