this post was submitted on 02 Sep 2024
88 points (100.0% liked)
Technology
37717 readers
737 users here now
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Where I don't think your argument fits is that it could be applied to things LLMs can currently do. If I have an insufficiently trained model which produces a word salad to every prompt, one could say "that's not a malfunction, it's still applying weights."
The malfunction is in having a system that produces useful results. An LLM is just the means for achieving that result, and you could argue it's the wrong tool for the job and that's fine. If I put gasoline in my diesel car and the engine dies, I can still say the car is malfunctioning. It's my fault, and the engine wasn't ever supposed to have gas in it, but the car is now "failing to function in a normal or satisfactory manner," the definition of malfunction.
The purpose of an LLM, at a fundamental level, is to approximate text it was trained on. If it was trained on gibberish, outputting gibberish wouldn't be a bug. If it wasn't, outputting gibberish would be indicative of a bug.
A better analogy would be selling someone a diesel car, when they wanted an electric vehicle, and them being upset when it requires refueling with gas. The car isn't malfunctioning in that case, the salesman was.
I'd argue that's what an LLM is, not its purpose. Continuing the car analogy, that's like saying a car's purpose is to burn gasoline to spin its wheels. That's what a car does, the purpose of my car is to get me from place to place. The purpose of my friend's car is to look cool and go fast. The purpose of my uncle's car is to carry lumber.
I think we more or less agree on the fundamentals and it's just differences between whether they are referring to a malfunction in the system they are trying to create, in which an LLM is a key tool/component, or a malfunction in the LLM itself. At the end of the day, I think we can all agree that it did a thing they didn't want it to do, and that an LLM by itself may not be the correct tool for the job.
No, that was the purpose for you, that made you choose to buy it. Someone else could have chosen to buy a car to live in it, for example. The purpose of a tool is just to be a tool. A hammer's purpose isn't just to hit nails with, it's to be a heavy thing you can use as-needed. You could hit a person with it, or straighten out dents in a metal sheet, or destroy a harddrive. I think you're conflating the intended use of something, with its purpose for existing, and it's leading you to assert that the purpose of LLMs is one specific use only.
An LLM is never going to be a fact-retrieval engine, but it has plenty of legitimate uses: generating creative text is very useful. Just because OpenAI is selling their creative-text engine under false pretenses doesn't invalidate the technology itself.
Sure, 100% they are using/ selling the wrong tool for the job, but the tool is not malfunctioning.