this post was submitted on 28 Jan 2025
876 points (94.5% liked)

memes

11300 readers
2566 users here now

Community rules

1. Be civilNo trolling, bigotry or other insulting / annoying behaviour

2. No politicsThis is non-politics community. For political memes please go to !politicalmemes@lemmy.world

3. No recent repostsCheck for reposts when posting a meme, you can only repost after 1 month

4. No botsNo bots without the express approval of the mods or the admins

5. No Spam/AdsNo advertisements or spam. This is an instance rule and the only way to live.

A collection of some classic Lemmy memes for your enjoyment

Sister communities

founded 2 years ago
MODERATORS
 

Office space meme:

"If y'all could stop calling an LLM "open source" just because they published the weights... that would be great."

you are viewing a single comment's thread
view the rest of the comments
[–] thespcicifcocean@lemmy.world 2 points 2 days ago (2 children)

It's not just the weights though is it? You can download the training data they used, and run your own instance of the model completely separate from their servers.

[–] BradleyUffner@lemmy.world 8 points 2 days ago* (last edited 2 days ago)

You don't download the training data when running an LLM locally. You are downloading the already baked model.

[–] Prunebutt@slrpnk.net 9 points 2 days ago (1 children)

Did "they" publish the training data? And the hyperparameters?

[–] thespcicifcocean@lemmy.world -2 points 2 days ago (1 children)

I mean, I downloaded it from the repo.

[–] Prunebutt@slrpnk.net 10 points 2 days ago (1 children)

You downloaded the weights. That's something different.

[–] thespcicifcocean@lemmy.world 1 points 2 days ago (1 children)

I may misunderstand, but are the weights typically several hundred gigabytes large?

[–] Prunebutt@slrpnk.net 8 points 2 days ago* (last edited 2 days ago) (1 children)

Yes. The training data is probably a few hundred petabytes.

[–] thespcicifcocean@lemmy.world 2 points 2 days ago (1 children)
[–] BradleyUffner@lemmy.world 3 points 2 days ago

Yeah, some models are trained on pretty much the entire content of the publicly accessible Internet.