About as open source as a binary blob without the training data
-
[email protected]replied to [email protected] last edited by
I may misunderstand, but are the weights typically several hundred gigabytes large?
-
[email protected]replied to [email protected] last edited by
Even worse is calling a proprietary, absolutely closed source, closed data and closed weight company "OpeanAI"
-
Is it even really software, or just a datablob with a dedicated interpreter?
-
[email protected]replied to [email protected] last edited by
That's fat shaming
-
[email protected]replied to [email protected] last edited by
For neural nets the method matters more. Data would be useful, but at the amount these things get trained on the specific data matters little.
They can be trained on anything, and a diverse enough data set would end up making it function more or less the same as a different but equally diverse set. Assuming publicly available data is in the set, there would also be overlap.
The training data is also by necessity going to be orders of magnitude larger than the model itself. Sharing becomes impractical at a certain point before you even factor in other issues.
-
[email protected]replied to [email protected] last edited by
I mean that's all a model is so....
Once again someone who doesn't understand anything about training or models is posting borderline misinformation about ai.Shocker
-
[email protected]replied to [email protected] last edited by
Hell, for all we know it could be full of classified data. I guess depending on what country you're in it definitely is full of classified data...
-
[email protected]replied to [email protected] last edited by
Yet another so-called AI evangelist accusing others of not understanding computer science if they don't want to worship their machine god.
-
[email protected]replied to [email protected] last edited by
You don't download the training data when running locally. You are downloading the sheet baked model.
-
[email protected]replied to [email protected] last edited by
Praise the Omnisiah! ... I'll see myself out.
-
[email protected]replied to [email protected] last edited by
While I completely agree with 90% of your comment, that first sentence is gross hyperbole. I have used a number of pieces of open source options that are are clearly better. 7zip is a perfect example. For over a decade it was vastly superior to anything else, open or closed. Even now it may be showing its age a bit, but it is still one of the best options.
But for the rest of your statement, I completely agree. And yes, CAD is a perfect example of the problems faced by open source. I made the mistake of thinking that I should start learning CAD with open source and then I wouldn't have to worry about getting locked into any of the closed source solutions. But Freecad is such a mess. I admit it has gotten drastically better over the last few years, but it still has serious issues. Don't get me wrong, I still 100% recommend that people learn it, but I push them towards a number of closed source options to start with. Freecad is for advanced users only. -
[email protected]replied to [email protected] last edited by
Do you think your comments here are implying an understanding of the tech?
-
[email protected]replied to [email protected] last edited by
Judging by OP’s salt in the comments, I’m guessing they might be an Nvidia investor. My condolences.
-
[email protected]replied to [email protected] last edited by
Especially after it was founded as a nonprofit with the mission to push open source AI as far and wide as possible to ensure a multipolar AI ecosystem, in turn ensuring AI keeping other AI in check so that AI would be respectful and prosocial.
-
[email protected]replied to [email protected] last edited by
A model can be represented only by its weights in the same way that a codebase can be represented only by its binary.
Training data is a closer analogue of source code than weights.
-
[email protected]replied to [email protected] last edited by
7-zip
VLC
OBS
Firefox did it only to mostly falter to Chrome but Chrome is largely Chromium which is open source.
Linux (superseded all the Unix, very severely curtailed Windows Server market)
Nearly all programming language tools (IDEs, Compilers, Interpreters)
Essentially all command line ecosystem (obviously on the *nix side, but MS was pretty much compelled to open source Powershell and their new Terminal to try to compete)In some contexts you aren't going to have a lively enough community to drive a compelling product even as there's enough revenue to facilitate a company to make a go of it, but to say 'no open source software has acheived that' is a bit much.
-
[email protected]replied to [email protected] last edited by
A model is an artifact, not the source. We also don't call binaries "open-source", even though they are literally the code that's executed. Why should these phrases suddenly get turned upside down for AI models?
-
Fushuan [he/him]replied to magic_lobster_party last edited by
Hey, I have trained several models in pytorch, darknet, tensorflow.
With the same dataset and the same training parameters, the same final iteration of training actually does return the same weights. There's no randomness unless they specifically add random layers and that's not really a good idea with RNNs it wasn't when I was working with them at least. In any case, weights should converge into a very similar point even if randomness is introduced or else the RNN is pretty much worthless.
-
magic_lobster_partyreplied to Fushuan [he/him] last edited by
There’s usually randomness involved with the initial weights and the order the data is processed.
-
Fushuan [he/him]replied to magic_lobster_party last edited by
Not enough for it to make results diverge. Randomness is added to avoid falling into local maximas in optimization. You should still end in the same global maxima. Models usualy run until their optimization converges.
As stated, if the randomness is big enough that multiple reruns end up with different weights aka optimized for different maximas, the randomization is trash. Anything worth their salt won't have randomization big enough.
So, going back to my initial point, we need the training data to validate the weights. There are ways to check the performance of a model (quite literally, the same algorithm that is used to evaluate weights in training is them used to evaluate the trained weights post training) the performance should be identical up to a very small rounding error if a rerun with the same data and parameters is used.