Difference between revisions of "LLaMA"

From
Jump to: navigation, search
m
m (Text replacement - "* Conversational AI ... ChatGPT | OpenAI ... Bing | Microsoft ... Bard | Google ... Claude | Anthropic ... Perplexity ... You ... Ernie | Baidu" to "* Conversational AI ... [[C...)
Line 25: Line 25:
 
* [[Toolformer]]
 
* [[Toolformer]]
 
* [[What is Artificial Intelligence (AI)? | Artificial Intelligence (AI)]] ... [[Generative AI]] ... [[Machine Learning (ML)]] ... [[Deep Learning]] ... [[Neural Network]] ... [[Reinforcement Learning (RL)|Reinforcement]] ... [[Learning Techniques]]
 
* [[What is Artificial Intelligence (AI)? | Artificial Intelligence (AI)]] ... [[Generative AI]] ... [[Machine Learning (ML)]] ... [[Deep Learning]] ... [[Neural Network]] ... [[Reinforcement Learning (RL)|Reinforcement]] ... [[Learning Techniques]]
* [[Conversational AI]] ... [[ChatGPT]] | [[OpenAI]] ... [[Bing]] | [[Microsoft]] ... [[Bard]] | [[Google]] ... [[Claude]] | [[Anthropic]] ... [[Perplexity]] ... [[You]] ... [[Ernie]] | [[Baidu]]
+
* [[Conversational AI]] ... [[ChatGPT]] | [[OpenAI]] ... [[Bing/Copilot]] | [[Microsoft]] ... [[Gemini]] | [[Google]] ... [[Claude]] | [[Anthropic]] ... [[Perplexity]] ... [[You]] ... [[phind]] ... [[Ernie]] | [[Baidu]]
 
* [[Assistants]] ... [[Personal Companions]] ... [[Agents]]  ... [[Negotiation]] ... [[LangChain]]
 
* [[Assistants]] ... [[Personal Companions]] ... [[Agents]]  ... [[Negotiation]] ... [[LangChain]]
 
* [https://arstechnica.com/information-technology/2023/02/chatgpt-on-your-pc-meta-unveils-new-ai-model-that-can-run-on-a-single-gpu/ Meta unveils a new large language model that can run on a single GPU | Benj Edwards - Ars Technica]  ... LLaMA-13B reportedly outperforms [[ChatGPT]]-like tech despite being 10x smaller.
 
* [https://arstechnica.com/information-technology/2023/02/chatgpt-on-your-pc-meta-unveils-new-ai-model-that-can-run-on-a-single-gpu/ Meta unveils a new large language model that can run on a single GPU | Benj Edwards - Ars Technica]  ... LLaMA-13B reportedly outperforms [[ChatGPT]]-like tech despite being 10x smaller.

Revision as of 10:25, 16 March 2024

YouTube search... ...Google search


LLaMA is a Large Language Model (LLM) released by Meta Platforms Inc. (formerly Facebook Inc.).

LLaMA 2 long

Llama 2 Long is a 137B parameter model that was trained on a massive dataset of text and code. It is able to generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way. On some benchmarks, Llama 2 Long outperforms GPT-3.5 Turbo and Claude 2 by a significant margin. For example, on the BLEU-4 benchmark, which measures the fluency of machine-generated text, Llama 2 Long scored 104.5, while GPT-3.5 Turbo scored 102.1 and Claude 2 scored 103.2. Meta is making Llama 2 Long available for free for research and commercial use. This is a significant step forward in the development of open source AI models, and it is likely to lead to new and innovative applications for large language models.

LLaMA 2

Meta released LLaMA in July 2023. The company said that it hopes that by making LLaMA 2 open-source, it will be able to improve the model by getting feedback from the wider community of developers. Microsoft and Meta are expanding their longstanding partnership, with Microsoft as the preferred partner for Llama 2. LLaMA is still under development, but it has already been used to create some impressive chatbots. For example, one chatbot called Allie can be used to provide customer support. Another chatbot called Galactica is designed for scientific research. LLaMA-2-7B, LLaMA-2-13B, and LLaMA-2-70B.


LLaMA (initial)

LLaMA on February 24, 20231. Meta says it is democratizing access to LLMs, which are seen as one of the most important and beneficial forms of AI. The four foundation models of LLaMA are LLaMA-7B, LLaMA-13B, LLaMA-33B, and LLaMA-65B12. They have 7 billion, 13 billion, 33 billion, and 65 billion parameters respectively. The models are all based on the transformer architecture and trained on publicly available datasets. LLaMA-13B is remarkable because it can run on a single GPU and outperform GPT-3 (175 billion parameters) on most common sense reasoning benchmarks. LLaMA-65B is competitive with the best models from other AI labs, such as Chinchilla 70B and PaLM 540B.