Grok AI Goes Open Source (venturebeat.com) 38
xAI has opened sourced its large language model Grok. From a report: The move, which Musk had previously proclaimed would happen this week, now enables any other entrepreneur, programmer, company, or individual to take Grok's weights -- the strength of connections between the model's artificial "neurons," or software modules that allow the model to make decisions and accept inputs and provide outputs in the form of text -- and other associated documentation and use a copy of the model for whatever they'd like, including for commercial applications.
"We are releasing the base model weights and network architecture of Grok-1, our large language model," the company announced in a blog post. "Grok-1 is a 314 billion parameter Mixture-of-Experts model trained from scratch by xAI." Those interested can download the code for Grok on its Github page or via a torrent link. Parameters refers to the weights and biases that govern the model -- the more parameters, generally the more advanced, complex and performant the model is. At 314 billion parameters, Grok is well ahead of open source competitors such as Meta's Llama 2 (70 billion parameters) and Mistral 8x7B (12 billion parameters). Grok was open sourced under an Apache License 2.0, which enables commercial use, modifications, and distribution, though it cannot be trademarked and there is no liability or warranty that users receive with it. In addition, they must reproduce the original license and copyright notice, and state the changes they've made.
"We are releasing the base model weights and network architecture of Grok-1, our large language model," the company announced in a blog post. "Grok-1 is a 314 billion parameter Mixture-of-Experts model trained from scratch by xAI." Those interested can download the code for Grok on its Github page or via a torrent link. Parameters refers to the weights and biases that govern the model -- the more parameters, generally the more advanced, complex and performant the model is. At 314 billion parameters, Grok is well ahead of open source competitors such as Meta's Llama 2 (70 billion parameters) and Mistral 8x7B (12 billion parameters). Grok was open sourced under an Apache License 2.0, which enables commercial use, modifications, and distribution, though it cannot be trademarked and there is no liability or warranty that users receive with it. In addition, they must reproduce the original license and copyright notice, and state the changes they've made.
Comment removed (Score:5, Informative)
Re: (Score:3)
GrokAI going open source is like (Score:5, Funny)
Re:GrokAI going open source is like (Score:5, Insightful)
Re: (Score:3)
Re: (Score:3)
I dunno, I'm going to wait for some actual benchmarks.
It's a gigantic model, by far the largest true-open-source (aka Apache, MIT, or similar licensed) model out there. But I strongly suspect it's severely undertrained relative to its size.
Would probably be a great base for any research on increasing information density in models via downscaling, though.
Comment removed (Score:4, Informative)
Re: (Score:3)
Re: (Score:3, Informative)
I think Linus said that Stranger in a Strange Land was his favourite book growing up. https://archive.computerhistor... [computerhistory.org]
(I wonder if it could be the origin of the name 'git')
Re: (Score:3, Funny)
Now, here you are, all grown up and still an idiot.
Re: (Score:2)
it was a good book.
Nope. From what I recall, it was overly dense and not fun to read because of it. Read it once, forgot most of it, and will never pick it up again.
Re: (Score:3)
What you get out of a book depends on what you bring to it. That said, it became popular because of the zeitgeist of the time. These days it would probably seem ... either silly or blindly utopian.
Re: (Score:1)
What you get out of a book depends on what you bring to it.
I bring nothing but the ability to read. It is up to the author to craft a good story.
The LotR books have the same issue. A lot of stuff before 1970 does honestly.
I swear, were people back in the day being paid by each unique word they put into a story?
Re: (Score:2)
Sure, maybe I am one of those 'pretentious asshats' that someone in this threa
Re: (Score:2)
there should be something to it that is good enough that people like to read it.
That, or it had good marketing.
Re: (Score:2)
Elon himself has made the same mistake ;)
The amount of AI news (Score:4, Insightful)
is fucking debilitating. Even Soviet propaganda didn't come that thick and fast.
Re: (Score:1)
Would you like some bitcoin stories instead?
Re:The amount of AI news (Score:4, Funny)
Man, you seem stressed. Why not vent to ChatGPT about your problems and ask it for some advice on how to deal with the situation?
Re: (Score:2)
Man, you seem stressed. Why not vent to ChatGPT about your problems and ask it for some advice on how to deal with the situation?
Oddly enough, I can see that as being a useful use of ChatGPT or another "AI" platform if it can be progammed to provide useful suggestions to people venting their problems at it. A FreudBot if you will. A lot of people will be reluctant to use mental health services, even if they're abundant and free, because they are scared of judgement when talking to an actual person. This is a much more surmountable barrier when talking to a machine.
OTOH, you'd want it to be managed and programmed by someone reputab
Re: (Score:1)
Be the ignore you want to see in the world.
Re: (Score:2)
Too big for casual use (Score:5, Informative)
The model is too big for casual use (IE running inference at home). LLama can just barely fit on either one 48 GB GPU, or two 24 GB GPUs. The number of parameters in Grok is much, much greater, really placing it outside the realistic realm of those wanting to run inference on their own hardware in a more hobbyist fashion. Sure, you can always run it on a normal CPU with a ton of RAM, but you'd be looking at multiple seconds per token.
I was going to ask... (Score:2)
Sure, "more" sounds better, but those need to live somewhere. And I haven't seen a GPU with an SSD yet to use as virtual memory (or whatever way you'd get crazy high amounts of parameters local).
Scaling by the same factor gives 215GB of VRAM necessary, right? And me still sitting here with a 1080 ti...
Hmm... 11GB dedicated, and 16GB shared? That's probably just CPU RAM that it can address (doesn't really speed things along).
Re: (Score:2)
Might this actually be a use case for Optane?
Might be interesting (Score:2)
Grok-1 doesn't seem like a terrible deal for half way between ChatGPT 3.5 and 4. 314B parameters, 8 experts, 2 experts per token... If that translates to something like 70B parameters per token mere mortals with lots of ram should be able to get a few tokens/s on a CPU.
I wonder how Groq has decided to play this. (Score:2)
They start the trademark claim yet?
Re: (Score:2)
On what basis?
The word 'grok' has existed since 1961, and has been a word of common colloquial use (in some demographics) for decades. It's in the New Hacker's Dictionary, descendant from The Jargon File (Jargon-1), which goes back to 1975. I was using the term in the 90s in high school.
Their misspelling is trademarked, not the word, which has clear prior art. (That's also probably why the model is called grok-1, I assume.)
As near as I can tell, there's no overlap between Groq and grok.
Furthermore, grok's e
How exactly does this work? (Score:2)
Can I run this locally somehow, ask it things and get the same results back as I would talking to the real Grok AI? Does this include all the training data?
Anyone? (Score:2)
The move ... now enables any other entrepreneur, programmer, company, or individual to take Grok's weights ... and use a copy of the model for whatever they'd like...
That should probably read "Any[one] who happens to not only have 0.65TB of spare storage but also has access to enormous amounts of GPU processing." The model will just about fit in an AWS P5 instance, with a current on-demand price of $98.32 per hour, or $860K per year before bulk discounts. I'm sure that there will be plenty of takers for this, but I think it's a bit of a stretch to say any other entrepreneur, programmer, company, or individual just yet.
Will Musk let it live.. (Score:2)
If it says anything even remotely negative about Musk?