this post was submitted on 15 May 2024
116 points (96.0% liked)

No Stupid Questions

35808 readers
1976 users here now

No such thing. Ask away!

!nostupidquestions is a community dedicated to being helpful and answering each others' questions on various topics.

The rules for posting and commenting, besides the rules defined here for lemmy.world, are as follows:

Rules (interactive)


Rule 1- All posts must be legitimate questions. All post titles must include a question.

All posts must be legitimate questions, and all post titles must include a question. Questions that are joke or trolling questions, memes, song lyrics as title, etc. are not allowed here. See Rule 6 for all exceptions.



Rule 2- Your question subject cannot be illegal or NSFW material.

Your question subject cannot be illegal or NSFW material. You will be warned first, banned second.



Rule 3- Do not seek mental, medical and professional help here.

Do not seek mental, medical and professional help here. Breaking this rule will not get you or your post removed, but it will put you at risk, and possibly in danger.



Rule 4- No self promotion or upvote-farming of any kind.

That's it.



Rule 5- No baiting or sealioning or promoting an agenda.

Questions which, instead of being of an innocuous nature, are specifically intended (based on reports and in the opinion of our crack moderation team) to bait users into ideological wars on charged political topics will be removed and the authors warned - or banned - depending on severity.



Rule 6- Regarding META posts and joke questions.

Provided it is about the community itself, you may post non-question posts using the [META] tag on your post title.

On fridays, you are allowed to post meme and troll questions, on the condition that it's in text format only, and conforms with our other rules. These posts MUST include the [NSQ Friday] tag in their title.

If you post a serious question on friday and are looking only for legitimate answers, then please include the [Serious] tag on your post. Irrelevant replies will then be removed by moderators.



Rule 7- You can't intentionally annoy, mock, or harass other members.

If you intentionally annoy, mock, harass, or discriminate against any individual member, you will be removed.

Likewise, if you are a member, sympathiser or a resemblant of a movement that is known to largely hate, mock, discriminate against, and/or want to take lives of a group of people, and you were provably vocal about your hate, then you will be banned on sight.



Rule 8- All comments should try to stay relevant to their parent content.



Rule 9- Reposts from other platforms are not allowed.

Let everyone have their own content.



Rule 10- Majority of bots aren't allowed to participate here.



Credits

Our breathtaking icon was bestowed upon us by @Cevilia!

The greatest banner of all time: by @TheOneWithTheHair!

founded 1 year ago
MODERATORS
 

I have many conversations with people about Large Language Models like ChatGPT and Copilot. The idea that "it makes convincing sentences, but it doesn't know what it's talking about" is a difficult concept to convey or wrap your head around. Because the sentences are so convincing.

Any good examples on how to explain this in simple terms?

Edit:some good answers already! I find especially that the emotional barrier is difficult to break. If an AI says something malicious, our brain immediatly jumps to "it has intent". How can we explain this away?

top 50 comments
sorted by: hot top controversial new old
[–] Tar_alcaran@sh.itjust.works 66 points 6 months ago

It's a really well-trained parrot. It responds to what you say, and then it responds to what it hears itself say.

But despite knowing which sounds go together based on which sounds it heard, it doesn't actually speak English.

[–] CodeInvasion@sh.itjust.works 47 points 6 months ago* (last edited 6 months ago) (2 children)

I am an LLM researcher at MIT, and hopefully this will help.

As others have answered, LLMs have only learned the ability to autocomplete given some input, known as the prompt. Functionally, the model is strictly predicting the probability of the next word^+^, called tokens, with some randomness injected so the output isn’t exactly the same for any given prompt.

The probability of the next word comes from what was in the model’s training data, in combination with a very complex mathematical method to compute the impact of all previous words with every other previous word and with the new predicted word, called self-attention, but you can think of this like a computed relatedness factor.

This relatedness factor is very computationally expensive and grows exponentially, so models are limited by how many previous words can be used to compute relatedness. This limitation is called the Context Window. The recent breakthroughs in LLMs come from the use of very large context windows to learn the relationships of as many words as possible.

This process of predicting the next word is repeated iteratively until a special stop token is generated, which tells the model go stop generating more words. So literally, the models builds entire responses one word at a time from left to right.

Because all future words are predicated on the previously stated words in either the prompt or subsequent generated words, it becomes impossible to apply even the most basic logical concepts, unless all the components required are present in the prompt or have somehow serendipitously been stated by the model in its generated response.

This is also why LLMs tend to work better when you ask them to work out all the steps of a problem instead of jumping to a conclusion, and why the best models tend to rely on extremely verbose answers to give you the simple piece of information you were looking for.

From this fundamental understanding, hopefully you can now reason the LLM limitations in factual understanding as well. For instance, if a given fact was never mentioned in the training data, or an answer simply doesn’t exist, the model will make it up, inferring the next most likely word to create a plausible sounding statement. Essentially, the model has been faking language understanding so much, that even when the model has no factual basis for an answer, it can easily trick a unwitting human into believing the answer to be correct.

—-

^+^more specifically these words are tokens which usually contain some smaller part of a word. For instance, understand and able would be represented as two tokens that when put together would become the word understandable.

[–] HamsterRage@lemmy.ca 9 points 6 months ago

I think that a good starting place to explain the concept to people would be to describe a Travesty Generator. I remember playing with one of those back in the 1980's. If you fed it a snippet of Shakespeare, what it churned out sounded remarkably like Shakespeare, even if it created brand "new" words.

The results were goofy, but fun because it still almost made sense.

The most disappointing source text I ever put in was TS Eliot. The output was just about as much rubbish as the original text.

[–] Sabata11792@kbin.social 4 points 6 months ago (1 children)

As some nerd playing with various Ai models at home with no formal training, any wisdom you think that's worth sharing?

[–] BigMikeInAustin@lemmy.world 9 points 6 months ago (1 children)

The only winning move is not to play.

[–] Sabata11792@kbin.social 5 points 6 months ago

But my therapist said she needs more VRam.

[–] IzzyScissor@lemmy.world 24 points 6 months ago

It's your phone's 'predictive text', but if it were trained on the internet.

It can guess what the next word should be a lot of the time, but it's also easy for it to go off the rails.

[–] GamingChairModel@lemmy.world 23 points 6 months ago

Harry Frankfurt's influential 2005 book (based on his influential 1986 essay), On Bullshit, offered a description of what bullshit is.

When we say a speaker tells the truth, that speaker says something true that they know is true.

When we say a speaker tells a lie, that speaker says something false that they know is false.

But bullshit is when the speaker says something to persuade, not caring whether the underlying statement is true or false. The goal is to persuade the listener of that underlying fact.

The current generation of AI chat bots are basically optimized for bullshit. The underlying algorithms reward the models for sounding convincing, not necessarily for being right.

[–] NeoNachtwaechter@lemmy.world 22 points 6 months ago (1 children)

idea that "it makes convincing sentences, but it doesn't know what it's talking about"

Like a teenager who has come into a new group and is now trying so hard to fit in :-)

[–] MeatsOfRage@lemmy.world 11 points 6 months ago (1 children)

I think we forget this, it's just doing what people do

[–] Drummyralf@lemmy.world 6 points 6 months ago (5 children)

I commented something similair on another post, but this is exactly why I find this phenomenon so hard to describe.

A teenager in a new group still has some understanding and has a mind. It knows many of the meaning of the words that are said. Sure, some catchphrases might be new, but general topics shouldn't be too hard to follow.

This is nothing like genAI. GenAI doesn't know anything at all. It has (simplified) a list of words that somehow are connected to eachother. But AI has no meaning of a wheel, what round is, what rolling is, what rubber is, what an axle is. NO understanding. Just words that happened to describe all of it. For us humans it is so difficult to understand that something uses language without knowing ANY of the meaning.

How can we describe this so our brains make sense that you can have language without understanding? The Chinese Room experiment comes close, but is quite complicated to explain as well I think.

load more comments (5 replies)
[–] Deconceptualist@lemm.ee 18 points 6 months ago (2 children)

You could maybe just share a meme like this one.

Some folks in the comments there share actual LLM results, a few of which are sensible but plenty that aren't far off from the joke.

[–] otp@sh.itjust.works 4 points 6 months ago (1 children)

LMAO! I tried it, and it said:

The man takes the goat across the river first, then he returns alone and takes the boat across again. Finally, he brings the goat's friend, Mr. Cabbage, across the river.

LMAO

I asked what if the man can't swim...

In that case, the man can still take the goat across the river first. Then, he ties the goat to the boat and pulls the boat across the river using a rope, ensuring he stays on the side of the river where he can safely manage the boat without swimming.

I asked who Mr. Cabbage is...

My apologies for the confusion! Mr. Cabbage is a playful addition. In the classic riddle, there's usually no mention of Mr. Cabbage. It's just the man, the goat, and perhaps another item or animal depending on the version of the riddle. I just added Mr. Cabbage for some extra fun!

Then I asked what some other additions could be...

Certainly! You can add various characters or items to make the riddle more entertaining. Here are a few ideas:

  1. Mr. Wolf: The man, the goat, and Mr. Wolf need to cross the river. But if the man leaves the wolf alone with the goat, the wolf will eat the goat.

  2. Ms. Hen: The man, the goat, and Ms. Hen need to cross the river. But if the man leaves the hen alone with the goat, the goat will eat the hen.

  3. Professor Fox: The man, the goat, and Professor Fox need to cross the river. But if the man leaves the fox alone with the goat, the fox will eat the goat.

  4. Princess Diamond: The man, the goat, and Princess Diamond's precious gem need to cross the river. But Princess Diamond has warned that if anyone other than herself touches the gem, they'll be cursed.

Feel free to mix and match or come up with your own additions!

And the "solution"...

Ah, that's a fun twist! Let's see:

  1. The man takes Mr. Cabbage across first, then returns alone.
  2. Next, he takes Ms. Hen across but brings Mr. Cabbage back.
  3. He then takes Princess Diamond's gem across, leaves it on the other side, and takes Mr. Wolf across.
  4. The man leaves Mr. Wolf with Ms. Hen on the other side and returns alone.
  5. He takes Princess Diamond across, but before he leaves, he swaps Mr. Cabbage with Princess Diamond's gem.
  6. The man brings Mr. Cabbage across, leaving Princess Diamond's gem with Mr. Wolf.
  7. Finally, he takes the goat across.

Everyone safely crosses the river without any mishaps!

load more comments (1 replies)
load more comments (1 replies)
[–] BlameThePeacock@lemmy.ca 16 points 6 months ago (2 children)

It's just fancy predictive text like while texting on your phone. It guesses what the next word should be for a lot more complex topics.

[–] kambusha@sh.itjust.works 4 points 6 months ago

This is the one I got from the house to get the kids to the park and then I can go to work and then I can go to work and get the rest of the day after that I can get it to you tomorrow morning to pick up the kids at the same time as well as well as well as well as well as well as well as well as well... I think my predictive text broke

load more comments (1 replies)
[–] HorseRabbit@lemmy.sdf.org 16 points 6 months ago (5 children)

Not an ELI5, sorry. I'm an AI PhD, and I want to push back against the premises a lil bit.

Why do you assume they don't know? Like what do you mean by "know"? Are you taking about conscious subjective experience? or consistency of output? or an internal world model?

There's lots of evidence to indicate they are not conscious, although they can exhibit theory of mind. Eg: https://arxiv.org/pdf/2308.08708.pdf

For consistency of output and internal world models, however, their is mounting evidence to suggest convergence on a shared representation of reality. Eg this paper published 2 days ago: https://arxiv.org/abs/2405.07987

The idea that these models are just stochastic parrots that only probabilisticly repeat their training data isn't correct, although it is often repeated online for some reason.

A little evidence that comes to my mind is this paper showing models can understand rare English grammatical structures even if those structures are deliberately withheld during training: https://arxiv.org/abs/2403.19827

[–] GamingChairModel@lemmy.world 22 points 6 months ago (1 children)

The idea that these models are just stochastic parrots that only probabilisticly repeat their training data isn't correct

I would argue that it is quite obviously correct, but that the interesting question is whether humans are in the same category (I would argue yes).

[–] HorseRabbit@lemmy.sdf.org 13 points 6 months ago* (last edited 6 months ago)

People sometimes act like the models can only reproduce their training data, which is what I'm saying is wrong. They do generalise.

During training the models are trained to predict the next word, but after training the network is always effectively interpolating between the training examples it has memorised. But this interpolation doesn't happen in text space but in a very high dimensional abstract semantic representation space, a 'concept space'.

Now imagine that you have memorised two paragraphs that occupy two points in concept space. And then you interpolate between them. This gives you a new point, potentially unseen during training, a new concept, that is in some ways analogous to the two paragraphs you memorised, but still fundamentally different, and potentially novel.

[–] Rhynoplaz@lemmy.world 17 points 6 months ago

You sound like a chatbot who's offended by it's intelligence being insulted.

[–] meteokr@community.adiquaints.moe 15 points 6 months ago

I really appreciate you linking studies about this topic, as finding this kind of research can be daunting. Those looks like really interesting reads.

load more comments (2 replies)
[–] magic_lobster_party@kbin.run 14 points 6 months ago (2 children)

There’s the Chinese Room argument, which is a bit related:

https://en.wikipedia.org/wiki/Chinese_room

[–] Asifall@lemmy.world 5 points 6 months ago

I always thought the Chinese Room argument was kinda silly. It’s predicated on the idea that humans have some unique capacity to understand the world that can’t be replicated by a syntactic system, but there is no attempt made to actually define this capacity.

The whole argument depends on our intuition that we think and know things in a way inanimate objects don’t. In other words, it’s a tautology to draw the conclusion that computers can’t think from the premise that computers can’t think.

[–] Rolando@lemmy.world 4 points 6 months ago

This is what I was going to point to. When I was in grad school, it was often referred to as the Symbol Gounding Problem. Basically it's a interdisciplinary research problem involving pragmatics, embodied cognition, and a bunch of others. The LLM people are now crashing into this research problem, and it's interesting to see how they react.

[–] patatahooligan@lemmy.world 14 points 6 months ago

Imagine you were asked to start speaking a new language, eg Chinese. Your brain happens to work quite differently to the rest of us. You have immense capabilities for memorization and computation but not much else. You can't really learn Chinese with this kind of mind, but you have an idea that plays right into your strengths. You will listen to millions of conversations by real Chinese speakers and mimic their patterns. You make notes like "when one person says A, the most common response by the other person is B", or "most often after someone says X, they follow it up with Y". So you go into conversations with Chinese speakers and just perform these patterns. It's all just sounds to you. You don't recognize words and you can't even tell from context what's happening. If you do that well enough you are technically speaking Chinese but you will never have any intent or understanding behind what you say. That's basically LLMs.

[–] LainTrain@lemmy.dbzer0.com 13 points 6 months ago (1 children)

That analogy is hard to come up with because the question of whether it even comprehends meaning requires first answering the unanswerable question of what meaning actually is and whether or not humans are also just spicy pattern predictors / autocompletes, since predicting patterns is like the whole point of evolving intelligence, being able to connect cause and effect in patterns and anticipate the future just helps with not starving. The line is far blurrier than most are willing to admit and ultimately hinges on our experience of sapience rather than being able to strictly define knowledge and meaning.

Instead it's far better to say that ML models are not sentient, they are like a very big brain that's switched off, but we can access it by stimulating it with a prompt.

[–] Drummyralf@lemmy.world 6 points 6 months ago (1 children)

Interesting thoughts! Now that I think about this, we as humans have a huge advantage by having not only language, but also sight, smell, hearing and taste. An LLM basically only has "language." We might not realize how much meaning we create through those other senses.

[–] CodeInvasion@sh.itjust.works 5 points 6 months ago

To add to this insight, there are many recent publications showing the dramatic improvements of adding another modality like vision to language models.

While this is my conjecture that is loosely supported by existing research, I personally believe that multimodality is the secret to understanding human intelligence.

[–] Chocrates@lemmy.world 13 points 6 months ago

A 5 year old repeating daddy's swear words without knowing what it is.

[–] rubin@lemmy.sdf.org 12 points 6 months ago (1 children)

Imagine that you have a random group of people waiting in line at your desk. You have each one read the prompt, and the response so far, and then add a word themself. Then they leave and the next person in line comes and does it.

This is why "why did you say ?" questions are nonsensical to AI. The code answering it is not the code that wrote it and there is no communication coordination or anything between the different word answerers.

load more comments (1 replies)
[–] Feathercrown@lemmy.world 11 points 6 months ago

The idea that “it makes convincing sentences, but it doesn’t know what it’s talking about” is a difficult concept to convey or wrap your head around.

I see the people you talk to aren't familiar with politicians?

[–] TheBananaKing@lemmy.world 10 points 6 months ago

Imagine making a whole chicken out of chicken-nugget goo.

It will look like a roast chicken. It will taste alarmingly like chicken. It absolutely will not be a roast chicken.

The sad thing is that humans do a hell of a lot of this, a hell of a lot of the time. Look how well a highschooler who hasn't actually read the book can churn out a book report. Flick through, soak up the flavour and texture of the thing, read the blurb on the back to see what it's about, keep in mind the bloated over-flowery language that teachers expect, and you can bullshit your way to an A.

Only problem is, you can't use the results for anything productive, which is what people try to use GenAI for.

[–] JamesStallion@sh.itjust.works 9 points 6 months ago (4 children)

The Chinese Room by Searle

load more comments (4 replies)
[–] kava@lemmy.world 9 points 6 months ago (1 children)

It's all just fancy statistics. It turns words into numbers. Then it finds patterns in those numbers. When you enter a prompt, it finds numbers that are similar and spits out an answer.

You can get into vectors and back propagation and blah blah blah but essentially it's a math formula. We call it AI but it's not fundamentally different than solving 2x + 4 = 10 for x.

[–] Drummyralf@lemmy.world 7 points 6 months ago* (last edited 6 months ago) (4 children)

After reading some of the comments and pondering this question myself, I think I may have thought of a good analogy that atleast helps me (even though I know fairly well how LLM's work)

An LLM is like a car on the road. It can follow all the rules, like breaking in front of a red light, turning, signaling etc. However, a car has NO understanding of any of the traffic rules it follows.

A car can even break those rules, even if its behaviour is intended (if you push the gas pedal at a red light, the car is not in the wrong because it doesn't KNOW the rules, it just acts on it).

Why this works for me is that when I give examples of human behaviour or animal behaviour, I automatically ascribe some sort of consciousness. An LLM has no conscious (as far as I know for now). This idea is exactly what I want to convey. If I think of a car and rules, it is obvious to me that a car has no concept of rules, but still is part of those rules somehow.

load more comments (4 replies)
[–] AbouBenAdhem@lemmy.world 7 points 6 months ago* (last edited 5 months ago)
  • Compression algorithms can reduce most written text to about 20–25% of its original size—implying that that’s the amount of actual unique information it contains, while the rest is predictable filler.

  • Empirical studies have found that chimps and human infants, when looking at test patterns, will ignore patterns that are too predictable or too unpredictable—with the sweet spot for maximizing attention being patterns that are about 80% predictable.

  • AI researchers have found that generating new text by predicting the most likely continuation of the given input results in text that sounds monotonous and obviously robotic. Through trial and error, they found that, instead of choosing the most likely result, choosing one with around an 80% likelihood threshold produces results judged most interesting and human-like.

The point being: AI has stumbled on a method of mimicking the presence of meaning by imitating the ratio of novelty to predictability that characterizes real human thought. But we know that the actual content of that novelty is randomly chosen, rather than being a deliberate message.

[–] kromem@lemmy.world 6 points 6 months ago* (last edited 6 months ago) (2 children)

So there's two different things to what you are asking.

(1) They don't know what (i.e. semantically) they are talking about.

This is probably not the case, and there's very good evidence over the past year in research papers and replicated projects that transformer models do pick up world models from the training data such that they are aware and integrating things at a more conceptual level.

For example, even a small toy GPT model trained only on chess moves builds an internal structure of the whole board and tracks "my pieces" and "opponent pieces."

(2) Why do they say dumb shit that's clearly wrong and don't know.

They aren't knowledge memorizers. They are very advanced pattern extenders.

Where the answer to a question is part of the pattern they can successfully extend, they get the answer correct. But if it isn't, they confabulate an answer in a similar way to stroke patients who don't know that they don't know the answer to something and make it up as they go along. Similar to stroke patients, you can even detect when this is happening with a similar approach (ask 10x and see how consistent the answer is or if it changes each time).

They aren't memorizing the information like a database. They are building ways to extend input into output in ways that match as much information as they can be fed. In this, they are beyond exceptional. But they've been kind of shoehorned into the initial tech demo usecase of "knowledgeable chatbot" which is a less than ideal use. The fact they were even good at information recall was a surprise to most researchers.

load more comments (2 replies)
[–] Hamartiogonic@sopuli.xyz 5 points 6 months ago* (last edited 6 months ago) (1 children)

All of this also touches upon an interesting topic. What it really means to understand something? Just because you know stuff and may even be able to apply it in flexible ways, does that count as understanding? I’m not a philosopher, so I don’t even know how to approach something like this.

Anyway, I think the main difference is the lack of personal experience about the real world. With LLMs, it’s all second hand knowledge. A human could memorize facts like how water circulates between rivers, lakes and clouds, and all of that information would be linked to personal experiences, which would shape the answer in many ways. An LLM doesn’t have such experiences.

Another thing would be reflecting on your experiences and knowledge. LLMs do none of that. They just speak whatever “pops in their mind”, whereas humans usually think before speaking… Well at least we are capable of doing that even though we may not always take advantage of this super power. Although, the output of an LLM can be monitored and abruptly deleted as soon as it crosses some line. It’s sort of like mimicking the thought processes you have inside your head before opening your mouth.

Example: Explain what it feels like to have an MRI taken of your head. If you haven’t actually experienced that yourself, you’ll have to rely on second hand information. In that case, the explanation will probably be a bit flimsy. Imagine you also read all the books, blog posts and and reddit comments about it, and you’re able to reconstruct a fancy explanation regardless.

This lack of experience may hurt the explanation a bit, but an LLM doesn’t have any experiences of anything in the real world. It has only second hand descriptions of all those experiences, and that will severely hurt all explanations and reasoning.

[–] trashgirlfriend@lemmy.world 9 points 6 months ago (1 children)

I feel like you're already not getting it and therefore giving too much credit to the LLM.

With LLMs it's not even about second hand knowledge, the concept of knowledge does not apply to LLMs at all, it's literally just about statistics, eg. what is the most likely next output after this token.

load more comments (1 replies)
[–] DarkThoughts@fedia.io 5 points 6 months ago

Someone once described is as T9 on steroids. It's like your mobile keyboard suggesting follow up words, just a lot more complex in size.

If an AI says something malicious, our brain immediatly jumps to "it has intent". How can we explain this away?

The more you understand the underlying concept of LLMs, the more the magic fades away. LLMs are certainly cool and can be fun but the hype around them seems very artificial and they're certainly not what I'd describe as "AI". To me, an AI would be something that actually has some form of consciousness, something that actually can form its own thoughts and learn things on its own through observation or experimentation. LLMs can't do any of those things. They're static and always wait for your input to even do anything. For text generation you can even just regenerate an answer to the same previous text and the replies can and will vary greatly. If they say something mean or malicious, it's simply because it is based on whatever they were trained on and whatever parameters they are following (like if you told them to roleplay a mean person).

[–] Hegar@kbin.social 5 points 6 months ago* (last edited 6 months ago)

Part of the problem is hyperactive agency detection - the same biological bug/feature that fuels belief in the divine.

If a twig snaps, it could be nothing or someone. If it's nothing and we react as if it was someone, no biggie. If it was someone and we react as if it was nothing, potential biggie. So our brains are bias towards assuming agency where there is none, to keep us alive.

[–] DaDragon@kbin.social 5 points 6 months ago

The short hand answer I’d try to give people is ‘it’s statistics’. Based on training data, there’s a certain chance of certain words being in proximity of each other. There’s no reasoning behind placement, other than whatever pattern is discernible from known situation.

[–] unreasonabro@lemmy.world 5 points 6 months ago

It's like talking to a republican!

[–] nothacking@discuss.tchncs.de 4 points 6 months ago

Like a kid trying very hard to sound like everyone else. "Eloquent bullshit generator"

[–] Ziggurat@sh.itjust.works 4 points 6 months ago (2 children)

have you played that game where everyone write a subjet and put it on a stack of paper, then everyone puts a verb on a different stack of paper, then everyone put an object on a third stack of paper, and you can even add a place or whatever on the next stack of paper. You end-up with fun sentences like A cat eat Kevin's brain on the beach. It's the kind of stuff (pre-)teen do to have a good laugh.

Chat GPT somehow works the same way, except that instead of having 10 paper in 5 stack, it has millions of paper in thousands of stack and depending on the "context" will choose which stack it draws paper from (To take an ELI5 analogy)

load more comments (2 replies)
load more comments
view more: next ›