Meta made its Llama 2 AI model open-source because ‘Zuck has balls,’ a former top Facebook engineer says::Meta CEO Mark Zuckerberg took a big risk by making its powerful AI model Llama 2 mostly open source, according to Replit CEO Amjad Masad.
It probably has more to do with the fuck up they did with accidentally creating a torrent for the original model’s weights allowing them to spread across the internet. Doesn’t really take “balls” to open source it after that and make it look like it was intentional. Still good that they did however rather than trying to use legal intimidation on anyone who used the leaked models.
It was already available for non-commercial use. The difference was that you had to submit a form and it was a slow roll-out.
IIRC you could apply as a credentialed academic researcher but not as a member of the general public, but i could be wrong about that.
They didn’t check credentials, but it was indeed for research purposes. Folks were getting access if they said they were a student or researcher.
It’s not open source, though.
Removed by mod
The whole point of open-source is to be able to recreate it yourself so you can make changes. This is freeware. Free-as-in-beer, not free-as-in-speech. Hell, with freeware I can use it for commercial purposes, it’s not even as free as that.
Removed by mod
Can I remake the model only including Creative Commons sourced training material?
Removed by mod
I thought I was only licensed for non-commercial use
Removed by mod
Calling ML models “Open Source” is already confused. Because they are not programs, but rather formats, they don’t come 1:1 with the source.
You can obtain a model and train it futher. Similliar how you can get JPEG file with permissive licence, edit it and share it. Having the GIMP/Photoshop project from the image was created from is helpful but not nessesary.
here’s core difference: the nature of ai-models is generative, but all layers in a .PSD file are inherently static.
better analogy would be rendering of a fractal — a limited subset of infinite possibilities, but to explore the rest of them you need both rules and data
That’s 1 of the 2 lies in the headline. The other appears in quotation marks.
On the other hand, this could mean “this software it totally crap, and the users will shoot us down from all sides, and he has the balls to publish it anyway, even if we will get sued into kingdom come.”
Was not too bad for Google to make android open source
Google had no choice, when they bought Android it was already GPL
Android is Apache 2, only the kernel is GPL
Really unfortunetly…
It was probably just to get all of the free development work the community has done. There are multiple engines designed around optimizing llama models specifically such as llama.cpp and exllama, and many other projects built around the architecture.
Facebook’s research division also has a pretty consistent track record of releasing things to the public rather than letting their research models rot.
people are still talking like this?
This is the best summary I could come up with:
The AI community has embraced the opportunity, giving Meta CEO Mark Zuckerberg his next potentially huge platform.
This wouldn’t have happened unless Zuckerberg was willing to take a big risk on Llama 2 possibly being used for nefarious purposes, according to a former top Facebook engineer.
“It takes a certain amount of guts to release an open-source language model, especially with political heat that Meta’s getting from that,” said Amjad Masad during a recent episode of the No Priors podcast.
Before that, he spent almost 3 years at Facebook where he helped create React Native and other popular software development tools.
During the No Priors podcast, Masad said he’s been surprised that Meta is the only major tech company so far to go the open-source route for AI models.
He compared this to Facebook’s Open Compute project, which designed data center hardware and made that available for anyone to use and contribute to.
The original article contains 512 words, the summary contains 153 words. Saved 70%. I’m a bot and I’m open source!
Apparently, the Zuck fucks
If the main risk is the model being used for something nefarious, like teaching terrorists how to make weapons, then can we PLEASE stop calling Zuck “bold” for doing it?
Not caring about moral consequences is not bold. It’s reckless and uncaring. Sure, the jackass who built his startup on it now lives up Zuck’s ass in thanks, but the rest of us should call it what it is.
I always dislike this arguement. AI isn’t magically going to give people capabilities they otherwise wouldn’t have. Everything that it can do is just automating tasks that could already be done by humans.
If you want to know how to make bombs there are literally articles online for it, you don’t have to have an AI. If you did, then there wouldn’t be terrorists already.