Difference between revisions of "Out-of-Distribution (OOD) Generalization"

From
Jump to: navigation, search
m (BPeat moved page Out-of-Distribution (OOD) to Out-of-Distribution (OOD) Generalization without leaving a redirect)
m
Line 1: Line 1:
[[Out-of-Distribution (OOD)]] generalization refers to the ability of a machine learning model to generalize to new data that comes from a different distribution than the training data. This is a challenging problem because the testing distribution is unknown and different from the training distribution. There are several methods for improving out-of-distribution generalization. According to a survey on the topic, existing methods can be categorized into three parts based on their positions in the whole learning pipeline: unsupervised representation learning, supervised model learning and optimization. Another approach to out-of-distribution generalization is via learning domain-invariant features or hypothesis-invariant features.
+
* [[Math for Intelligence#Mathematical Reasoning|Mathematical Reasoning]]
 +
* [https://arxiv.org/abs/2108.13624 Towards Out-Of-Distribution Generalization: A Survey]
 +
* [https://arxiv.org/abs/2106.04496 Towards a Theoretical Framework of Out-of-Distribution Generalization]
 +
* [http://proceedings.mlr.press/v139/krueger21a/krueger21a.pdf Out-of-Distribution Generalization via Risk Extrapolation]
 +
* [https://arxiv.org/abs/2210.10636 Using Interventions to Improve Out-of-Distribution Generalization of ....
 +
* [https://link.springer.com/chapter/10.1007/978-3-030-92659-5_39 How Reliable Are Out-of-Distribution Generalization Methods for Medical ....
 +
* [https://link.springer.com/chapter/10.1007/978-3-031-25075-0_36 Meta-Causal Feature Learning for Out-of-Distribution Generalization ...]
 +
 
 +
Out-of-Distribution (OOD) generalization refers to the ability of a machine learning model to generalize to new data that comes from a different distribution than the training data. This is a challenging problem because the testing distribution is unknown and different from the training distribution. There are several methods for improving out-of-distribution generalization. According to a survey on the topic, existing methods can be categorized into three parts based on their positions in the whole learning pipeline: unsupervised representation learning, supervised model learning and optimization. Another approach to out-of-distribution generalization is via learning domain-invariant features or hypothesis-invariant features.
  
 
Source: Conversation with Bing, 5/27/2023
 
Source: Conversation with Bing, 5/27/2023

Revision as of 07:15, 27 May 2023

Out-of-Distribution (OOD) generalization refers to the ability of a machine learning model to generalize to new data that comes from a different distribution than the training data. This is a challenging problem because the testing distribution is unknown and different from the training distribution. There are several methods for improving out-of-distribution generalization. According to a survey on the topic, existing methods can be categorized into three parts based on their positions in the whole learning pipeline: unsupervised representation learning, supervised model learning and optimization. Another approach to out-of-distribution generalization is via learning domain-invariant features or hypothesis-invariant features.

Source: Conversation with Bing, 5/27/2023 (1) Teaching Algorithmic Reasoning via In-context Learning - arXiv.org. https://arxiv.org/pdf/2211.09066.pdf. (2) Algorithmic prompting or how to teach math to a large language model. https://the-decoder.com/how-to-teach-math-to-a-large-language-model/. (3) 7 Examples of Algorithms in Everyday Life for Students. https://www.learning.com/blog/7-examples-of-algorithms-in-everyday-life-for-students/. (4) How to write the Algorithm step by step? - Programming-point. http://programming-point.com/algorithm-step-by-step/.