this post was submitted on 05 Nov 2024
7 points (70.6% liked)

Free Open-Source Artificial Intelligence

2889 readers
1 users here now

Welcome to Free Open-Source Artificial Intelligence!

We are a community dedicated to forwarding the availability and access to:

Free Open Source Artificial Intelligence (F.O.S.A.I.)

More AI Communities

LLM Leaderboards

Developer Resources

GitHub Projects

FOSAI Time Capsule

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] hendrik@palaver.p3x.de 6 points 1 week ago* (last edited 1 week ago) (7 children)

Silly license. Can be used worldwide, just not within the European Union where I live... (But it's the same with Meta's most recent models. The Llama 3.2 usage policy also contains a clause like that.)

We should really get some proper AI policy out.

[–] stuner@lemmy.world 7 points 1 week ago (1 children)

i.e. it's most definitely not open source.

[–] hendrik@palaver.p3x.de 5 points 1 week ago* (last edited 1 week ago) (2 children)

Practically none of the open source AI models are open source. At least not in the sense that term is used for software. Some people try to apply the word to AI models or just use it as a buzzword. It doesn't mean you get the source to recreate it (the dataset in this case). And they also restrict use in different ways. Open Source in the AI world just means you're able to download the weights and do inference on your own hardware. And you can do it with this model. Yet the license contains quite some limitations. I think we should stop using the term open source for AI before it loses all it's meaning.

That doesn't mean they're all licensed the same. Some are licensed under a proper free software license and while you usually still don't get the dataset, you get all the freedoms to use/run, share and modify the models to your liking.

[–] stuner@lemmy.world 4 points 1 week ago (1 children)

IMHO the OSI is right, the designation "open source" should be reserved for those models that are actually open source (including training data). And apparently there are a few models that actually meet this criterion: "Though none are confirmed, the handful of models that Bdeir told MIT Technology Review are expected to land on the list are relatively small names, including Pythia by Eleuther, OLMo by Ai2, and models by the open-source collective LLM360." (https://www.technologyreview.com/2024/08/22/1097224/we-finally-have-a-definition-for-open-source-ai/)

Perhaps it would also be useful to have a name for models that release their weights under an OSI license, maybe "open weight"? However, this model would not even meet that... (same for Llama).

[–] hendrik@palaver.p3x.de 1 points 1 week ago* (last edited 1 week ago)

Perhaps it would also be useful to have a name for models that release their weights [...]

open-weight?

I think the companies mostly stopped releasing the training data after a lot of them got sued for copyright infringement. I believe Meta's first LLaMA still came with a complete list of datasets that went in. And I forgot the name if the project but the community actually recreated it due to the licensing of the official model at that time that only allowed research. But things changed since then. Meta opened up a lot. Training got more extensive and is still prohibitively expensive (maybe even more so). And the landscape got riddled with legal issues, compared to the very early days where it was mostly research with less attention by everyone.

[–] Smorty@lemmy.blahaj.zone 1 points 1 week ago (1 children)

So would the granite models count as "open source"? They do publish the training data they used.

[–] hendrik@palaver.p3x.de 2 points 1 week ago* (last edited 1 week ago)

Seems they've outlined the used datasets in Annex B of their paper. I haven't checked if the list is exhaustive and if the training code and scripts to prepare the data are there... If they are, I'd say this is indeed a proper open-source model. And the weights are licensed under an Apache license.

load more comments (5 replies)