Meta To Release Open-Source Commercial AI Model To Compete With OpenAI, Google 16
An anonymous reader quotes a report from ZDNet: Meta, formerly known as Facebook, is set to release a commercial version of LLaMA, its open-source large language model (LLM) that uses artificial intelligence (AI) to generate text, images, and code. LLaMA, which stands for Large Language Model Meta AI, was publicly announced in February as a small foundational model, and made available to researchers and academics. Now, the Financial Times is reporting that Meta is prepared to release the commercial version of the model, which would enable developers and businesses to build applications using the foundational model.
Since it's an open-source AI technology, commercial access to LLaMA gives businesses of all sizes the opportunity to adapt and improve the AI, accelerating technological innovation across various sectors and potentially leading to more robust models. Meta's LLaMA is available in 7, 13, 33, and 65 billion parameters, compared to ChatGPT's LLM, GPT-3.5, which has been confirmed to have 175 billion parameters. OpenAI hasn't said how many parameters GPT-4 has, but it's estimated to have over 1 trillion parameters -- the more parameters, the better the model can understand input and generate appropriate output.
Though open-source AI models already exist, launching Meta's LLaMA commercially is still a significant step, due to it being larger than many of the available open-source LLMs on the market, and the fact that it is from one of the biggest tech companies in the world. The launch means Meta is directly competing with Microsoft-backed OpenAI and Google, and that competition could mean significant advancements in the AI field. Closed or proprietary software, like that used in OpenAI's ChatGPT, has drawn criticism over transparency and security.
Since it's an open-source AI technology, commercial access to LLaMA gives businesses of all sizes the opportunity to adapt and improve the AI, accelerating technological innovation across various sectors and potentially leading to more robust models. Meta's LLaMA is available in 7, 13, 33, and 65 billion parameters, compared to ChatGPT's LLM, GPT-3.5, which has been confirmed to have 175 billion parameters. OpenAI hasn't said how many parameters GPT-4 has, but it's estimated to have over 1 trillion parameters -- the more parameters, the better the model can understand input and generate appropriate output.
Though open-source AI models already exist, launching Meta's LLaMA commercially is still a significant step, due to it being larger than many of the available open-source LLMs on the market, and the fact that it is from one of the biggest tech companies in the world. The launch means Meta is directly competing with Microsoft-backed OpenAI and Google, and that competition could mean significant advancements in the AI field. Closed or proprietary software, like that used in OpenAI's ChatGPT, has drawn criticism over transparency and security.
"understand" (Score:2)
OMg, there's the magic word "understand" being used again.
Re: (Score:2)
What is really odd is that we don't understand how nature made the jump from simple neural nets to things with consciousness. We're just a bunch of inputs and outputs connected to a mass of specialized nerve cells... it's not like there's a soul in there doing supernatural things, we can see similar awareness in other animals to varying degrees.
As best as I can tell the difference between us and the current AI is ongoing feedback, the ability to affect the inputs with the outputs, and a few orders of magni
Re: (Score:2)
*statistical regularities
**correspo
That and "appropriate output". (Score:2)
there's the magic word "understand" being used again.
Yup, indeed. And then there's the bit after that:
Nope. The correct would be "and generate better sounding output.".
LLMs are good a putting nice sounding words together, given the context.
Whether those words are actually an appropriate or completely made up answer is entirely outside the scope of LLMs - their scope is merely "most likely words, given the training dataset".
When a user asks a question, there is no vi
hurry up (Score:1)
because by the time the lawsuits are finished over training data these tech companies are gonna wish they stuck with Eliza.
Llama (Score:2)
Say, is Jeff Minter still alive?
Late (Score:2)
LLaMA was leaked by researchers three months ago and 4chan users have been going nuts with excellent finetunes ever since. They're gonna have to do better than stock LLaMA.
Re: (Score:2)
Re: (Score:2)
Legal and illegal usage of leaked weights don't matter that much when Meta's model is trained in a manner that could very well be stealing. This is still all a grey area. Before you disagree, remember that you very probably supported the free usage of DeCSS keys and yelled "It's just math! It's just a number!"
Re: (Score:2)
Re: (Score:3)
Re: (Score:1)
Will there be a 4-bit (or less) versions? (Score:2)