this post was submitted on 09 Aug 2023
-4 points (40.0% liked)

Science Fiction

13602 readers
2 users here now

Welcome to /c/ScienceFiction

December book club canceled. Short stories instead!

We are a community for discussing all things Science Fiction. We want this to be a place for members to discuss and share everything they love about Science Fiction, whether that be books, movies, TV shows and more. Please feel free to take part and help our community grow.

  1. Be civil: disagreements happen, but that doesn’t provide the right to personally insult others.
  2. Posts or comments that are homophobic, transphobic, racist, sexist, ableist, or advocating violence will be removed.
  3. Spam, self promotion, trolling, and bots are not allowed
  4. Put (Spoilers) in the title of your post if you anticipate spoilers.
  5. Please use spoiler tags whenever commenting a spoiler in a non-spoiler thread.

Lemmy World Rules

founded 1 year ago
MODERATORS
 

Who needs human when you have AI :p

you are viewing a single comment's thread
view the rest of the comments
[–] DrJenkem@lemmy.blugatch.tube 1 points 1 year ago (1 children)

Sure there may be a lot, but it's still finite. And already, social media is being filled with AI generated content. If the trend continues, human generated content will be dwarfed by AI generated content. And it's not going to be a simple process to distinguish between the two.

[–] FaceDeer@kbin.social 1 points 1 year ago (1 children)

Infinite training data isn't required.

It's actually fine to include some AI-generated data in your training set, the reason "model collapse" happens is when you train on only AI-generated content and you end up losing out on some of the less-common outputs. Without the less-common cases in the training data each generation of AI has less diverse information to learn from. If you make sure the training set is diverse enough then it should be fine.

All else fails, just make sure a lot of your data is from before 2023.

[–] DrJenkem@lemmy.blugatch.tube 1 points 1 year ago* (last edited 1 year ago) (1 children)

I think you misunderstand the problem. Sure it starts with small amounts of output fed into the input, but as it continues to generate large amounts of output, overtime, more and more of the output makes it into the input.

And again, limiting LLMs to pre-2023 training data ensures they never get smarter. Human knowledge expands as LLMs at best are locked into a constant state of 2023 knowledge.

[–] FaceDeer@kbin.social 1 points 1 year ago (1 children)

Sure it starts with small amounts of output fed into the input, but as it continues to generate large amounts of output, overtime, more and more of the output makes it into the input.

Not inevitably. You're assuming that each "generation" of AI is being trained on a data set that's just blindly harvested. AI trainers are already spending a huge amount of effort curating their training sets, it's become quite apparent that the quality of the training set is important and you can't just dump a giant raw pile of everything into it to get good results. This would just be another thing for them to consider.

[–] DrJenkem@lemmy.blugatch.tube 1 points 1 year ago (1 children)

To a certain extent, yes, the training data is blindly being dumped in. There's no way terabytes of training data is being manually reviewed for accuracy. If for no other reason, it doesn't economically make sense to do so. It's simply not feasible for humans to manually currate all of that data and even if they did, human error still exists.

[–] FaceDeer@kbin.social 1 points 1 year ago

Your disbelief doesn't mean it's not happening. The data sources that go into AIs are indeed curated selectively. Honestly, what do you think happens, a webcrawler is told to just "go nuts" and whatever random data it spits out gets fed right in? Trainers pick their sources carefully. They deduplicate it, they format it, they do a lot of work on it.

Perfection is not required. Human error is fine in manageable amounts.