Difference between revisions of "LLaMA"

From
Jump to: navigation, search
m
m
Line 14: Line 14:
 
* [[Toolformer]]
 
* [[Toolformer]]
 
* [[Generative AI]]  ... [[Conversational AI]] ... [[OpenAI]]'s [[ChatGPT]] ... [[Perplexity]]  ... [[Microsoft]]'s [[Bing]] ... [[You]] ...[[Google]]'s [[Bard]] ... [[Baidu]]'s [[Ernie]]
 
* [[Generative AI]]  ... [[Conversational AI]] ... [[OpenAI]]'s [[ChatGPT]] ... [[Perplexity]]  ... [[Microsoft]]'s [[Bing]] ... [[You]] ...[[Google]]'s [[Bard]] ... [[Baidu]]'s [[Ernie]]
* [[Assistants]] ... [[Hybrid Assistants]]  ... [[Agents]]  ... [[Negotiation]] ... [[Hugging_Face#HuggingGPT|HuggingGPT]] ... [[LangChain]]
+
* [[Assistants]]  ... [[Agents]]  ... [[Negotiation]] ... [[Hugging_Face#HuggingGPT|HuggingGPT]] ... [[LangChain]]
 
* [https://arstechnica.com/information-technology/2023/02/chatgpt-on-your-pc-meta-unveils-new-ai-model-that-can-run-on-a-single-gpu/ Meta unveils a new large language model that can run on a single GPU | Benj Edwards - Ars Technica]  ... LLaMA-13B reportedly outperforms [[ChatGPT]]-like tech despite being 10x smaller.
 
* [https://arstechnica.com/information-technology/2023/02/chatgpt-on-your-pc-meta-unveils-new-ai-model-that-can-run-on-a-single-gpu/ Meta unveils a new large language model that can run on a single GPU | Benj Edwards - Ars Technica]  ... LLaMA-13B reportedly outperforms [[ChatGPT]]-like tech despite being 10x smaller.
 
* [https://www.reuters.com/technology/meta-launch-ai-language-model-llama-2023-02-24/ Meta heats up Big Tech's AI arms race with new language model | Yuvraj Malik and Katie Paul - Reuters]
 
* [https://www.reuters.com/technology/meta-launch-ai-language-model-llama-2023-02-24/ Meta heats up Big Tech's AI arms race with new language model | Yuvraj Malik and Katie Paul - Reuters]

Revision as of 15:51, 20 April 2023

YouTube search... ...Google search


LLaMA is a Large Language Model (LLM) released by Meta Platforms Inc. (formerly Facebook Inc.) on February 24, 20231. Meta says it is democratizing access to LLMs, which are seen as one of the most important and beneficial forms of AI. The four foundation models of LLaMA are LLaMA-7B, LLaMA-13B, LLaMA-33B, and LLaMA-65B12. They have 7 billion, 13 billion, 33 billion, and 65 billion parameters respectively. The models are all based on the transformer architecture and trained on publicly available datasets. LLaMA-13B is remarkable because it can run on a single GPU and outperform GPT-3 (175 billion parameters) on most common sense reasoning benchmarks. LLaMA-65B is competitive with the best models from other AI labs, such as Chinchilla 70B and PaLM 540B.