Difference between revisions of "FLAN-T5 LLM"

From
Jump to: navigation, search
m (Text replacement - "* Conversational AI ... ChatGPT | OpenAI ... Bing | Microsoft ... Bard | Google ... Claude | Anthropic ... Perplexity ... You ... Ernie | Baidu" to "* Conversational AI ... [[C...)
m
 
Line 20: Line 20:
 
* [[What is Artificial Intelligence (AI)? | Artificial Intelligence (AI)]] ... [[Generative AI]] ... [[Machine Learning (ML)]] ... [[Deep Learning]] ... [[Neural Network]] ... [[Reinforcement Learning (RL)|Reinforcement]] ... [[Learning Techniques]]
 
* [[What is Artificial Intelligence (AI)? | Artificial Intelligence (AI)]] ... [[Generative AI]] ... [[Machine Learning (ML)]] ... [[Deep Learning]] ... [[Neural Network]] ... [[Reinforcement Learning (RL)|Reinforcement]] ... [[Learning Techniques]]
 
* [[Artificial General Intelligence (AGI) to Singularity]] ... [[Inside Out - Curious Optimistic Reasoning| Curious Reasoning]] ... [[Emergence]] ... [[Moonshots]] ... [[Explainable / Interpretable AI|Explainable AI]] ...  [[Algorithm Administration#Automated Learning|Automated Learning]]
 
* [[Artificial General Intelligence (AGI) to Singularity]] ... [[Inside Out - Curious Optimistic Reasoning| Curious Reasoning]] ... [[Emergence]] ... [[Moonshots]] ... [[Explainable / Interpretable AI|Explainable AI]] ...  [[Algorithm Administration#Automated Learning|Automated Learning]]
* [[Transfer Learning]]
+
* [[Perspective]] ... [[Context]] ... [[In-Context Learning (ICL)]] ... [[Transfer Learning]] ... [[Out-of-Distribution (OOD) Generalization]]
 +
* [[Causation vs. Correlation]] ... [[Autocorrelation]] ...[[Convolution vs. Cross-Correlation (Autocorrelation)]]  
 
* [[Large Language Model (LLM)]] ... [[Natural Language Processing (NLP)]]  ...[[Natural Language Generation (NLG)|Generation]] ... [[Natural Language Classification (NLC)|Classification]] ...  [[Natural Language Processing (NLP)#Natural Language Understanding (NLU)|Understanding]] ... [[Language Translation|Translation]] ...  [[Natural Language Tools & Services|Tools & Services]]
 
* [[Large Language Model (LLM)]] ... [[Natural Language Processing (NLP)]]  ...[[Natural Language Generation (NLG)|Generation]] ... [[Natural Language Classification (NLC)|Classification]] ...  [[Natural Language Processing (NLP)#Natural Language Understanding (NLU)|Understanding]] ... [[Language Translation|Translation]] ...  [[Natural Language Tools & Services|Tools & Services]]
 
* [[BLIP-2]] | Salesforce Research ... a generic and efficient pre-training strategy that bootstraps vision-language pre-training from off-the-shelf frozen pre-trained image encoders and frozen large language models. It achieves state-of-the-art performance on various vision-language tasks, despite having significantly fewer trainable parameters than existing methods.
 
* [[BLIP-2]] | Salesforce Research ... a generic and efficient pre-training strategy that bootstraps vision-language pre-training from off-the-shelf frozen pre-trained image encoders and frozen large language models. It achieves state-of-the-art performance on various vision-language tasks, despite having significantly fewer trainable parameters than existing methods.

Latest revision as of 16:42, 28 April 2024

YouTube search... ...Google search

T5 stands for “Text-To-Text Transfer Transformer” . It is a model developed by Google Research that converts every language problem into a text-to-text format. T5 is an encoder-decoder model pre-trained on a multi-task mixture of unsupervised and supervised tasks and for which each task is converted into a text-to-text format. It was presented in a paper by Google called "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer". T5 works well on a variety of tasks out-of-the-box by prepending a different prefix to the input corresponding to each task. Large Language Model (LLM) are AI models that have been trained on large amounts of text data and can generate human-like text. - T5 | Hugging Face

FLAN-T5 is an improved version of T5 with some architectural tweaks. FLAN stands for “Fine-tuned LAnguage Net”. It was developed by Google Research and is pre-trained on C4 only without mixing in the supervised tasks. FLAN-T5 is designed to be highly customizable, allowing developers to fine-tune it to meet their specific needs. This means that developers can adjust the model’s parameters and architecture to better fit the data and task at hand. This can result in improved performance and accuracy on specific tasks. For example, a developer could fine-tune FLAN-T5 on a specific dataset to improve its performance on a particular language translation task. This flexibility makes FLAN-T5 a powerful tool for Natural Language Processing (NLP) tasks.

FLAN-T5-XL ... FLAN-T5-XXL is 11b | Hugging Face