Difference between revisions of "Out-of-Distribution (OOD) Generalization"

From
Jump to: navigation, search
m
m
Line 11: Line 11:
 
[https://www.bing.com/news/search?q=ai+Out+Distribution+OOD+Generalization&qft=interval%3d%228%22 ...Bing News]
 
[https://www.bing.com/news/search?q=ai+Out+Distribution+OOD+Generalization&qft=interval%3d%228%22 ...Bing News]
  
* [[In-Context Learning (ICL)]] ... [[Context]] ... [[Causation vs. Correlation]] ... [[Out-of-Distribution (OOD) Generalization]] ... [[Transfer Learning]]
+
* [[In-Context Learning (ICL)]] ... [[Context]] ... [[Causation vs. Correlation]] ... [[Autocorrelation]] ... [[Out-of-Distribution (OOD) Generalization]] ... [[Transfer Learning]]
 
* [[Singularity]] ... [[Artificial Consciousness / Sentience|Sentience]] ... [[Artificial General Intelligence (AGI)| AGI]] ... [[Inside Out - Curious Optimistic Reasoning| Curious Reasoning]] ... [[Emergence]] ... [[Moonshots]] ... [[Explainable / Interpretable AI|Explainable AI]] ...  [[Algorithm Administration#Automated Learning|Automated Learning]]
 
* [[Singularity]] ... [[Artificial Consciousness / Sentience|Sentience]] ... [[Artificial General Intelligence (AGI)| AGI]] ... [[Inside Out - Curious Optimistic Reasoning| Curious Reasoning]] ... [[Emergence]] ... [[Moonshots]] ... [[Explainable / Interpretable AI|Explainable AI]] ...  [[Algorithm Administration#Automated Learning|Automated Learning]]
 
* [[Math for Intelligence#Mathematical Reasoning|Mathematical Reasoning]]
 
* [[Math for Intelligence#Mathematical Reasoning|Mathematical Reasoning]]

Revision as of 17:30, 27 May 2023

YouTube ... Quora ...Google search ...Google News ...Bing News

Out-of-Distribution (OOD) generalization refers to the ability of a machine learning model to generalize to new data that comes from a different distribution than the training data. This is a challenging problem because the testing distribution is unknown and different from the training distribution. There are several methods for improving out-of-distribution generalization. According to a survey on the topic, existing methods can be categorized into three parts based on their positions in the whole learning pipeline: unsupervised representation learning, supervised model learning and optimization. Another approach to out-of-distribution generalization is via learning domain-invariant features or hypothesis-invariant features.

Difference Between In-Context Learning and OOD Generalization

In-Context Learning (ICL) refers to the ability of a machine learning model to learn from a few examples provided in the context of a task, without any fine-tuning. This is also known as few-shot learning or zero-shot learning.

Out-of-distribution (OOD) generalization, on the other hand, refers to the ability of a machine learning model to generalize to new data that comes from a different distribution than the training data1.

The main difference between In-Context Learning (ICL) and OOD generalization is that in-context learning focuses on the ability of a model to learn from a few examples provided in the context of a task, while OOD generalization focuses on the ability of a model to generalize to new data that comes from a different distribution than the training data.