this post was submitted on 28 Oct 2024
122 points (83.5% liked)
Technology
59357 readers
3460 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
We're not talking about a "style", we're talking about producing finished work. The image generation models aren't style guides, they output final images which are produced from the ingestion of other images as training data. The source material might be actual art (or not) but it is generally the product of a real person (because ML ingesting its own products is very much a garbage-in garbage-out system) who is typically not compensated for their work. So again, these generative ML models are ripoff systems, and nothing more. And no, typing in a prompt doesn't count as innovation or creativity.
Generative ai is not only prompting, which shows you don't know. Who are you to decide what is creativity and innovation? Are you Mr Art?
Anyway, it is not ingesting images and photobashing them into a final picture, that's not how it works. It has no memory of training data images, instead it learned to generate images by trying and when similar to a training data image going more in that direction. So it has the ability to create in the same style, but the original images it doesn't have them
I see, so your argument is that because the training data is not stored in the model in its original form, it doesn't count as a copy, and therefore it doesn't constitute intellectual property theft. I had never really understood what the justification for this point of view was, so thanks for that, it's a bit clearer now. It's still wrong, but at least it makes some kind of sense.
If the model "has no memory of training data images", then what effect is it that the images have on the model? Why is the training data necessary, what is its function?
Here's a video explaining how diffusion models work, and this article by Kit Walsh, a senior staff attorney at the EFF.