Difference between revisions of "Out-of-Distribution (OOD) Generalization"

From
Jump to: navigation, search
m
m
Line 18: Line 18:
 
* [https://arxiv.org/abs/2106.04496 Towards a Theoretical Framework of Out-of-Distribution Generalization]
 
* [https://arxiv.org/abs/2106.04496 Towards a Theoretical Framework of Out-of-Distribution Generalization]
 
* [http://proceedings.mlr.press/v139/krueger21a/krueger21a.pdf Out-of-Distribution Generalization via Risk Extrapolation]
 
* [http://proceedings.mlr.press/v139/krueger21a/krueger21a.pdf Out-of-Distribution Generalization via Risk Extrapolation]
* [https://arxiv.org/abs/2210.10636 Using Interventions to Improve Out-of-Distribution Generalization of ....
+
* [https://arxiv.org/abs/2210.10636 Using Interventions to Improve Out-of-Distribution Generalization of ....]
* [https://link.springer.com/chapter/10.1007/978-3-030-92659-5_39 How Reliable Are Out-of-Distribution Generalization Methods for Medical ....
+
* [https://link.springer.com/chapter/10.1007/978-3-030-92659-5_39 How Reliable Are Out-of-Distribution Generalization Methods for Medical ....]
 
* [https://link.springer.com/chapter/10.1007/978-3-031-25075-0_36 Meta-Causal Feature Learning for Out-of-Distribution Generalization ...]
 
* [https://link.springer.com/chapter/10.1007/978-3-031-25075-0_36 Meta-Causal Feature Learning for Out-of-Distribution Generalization ...]
  
 
Out-of-Distribution (OOD) generalization refers to the ability of a machine learning model to generalize to new data that comes from a different distribution than the training data. This is a challenging problem because the testing distribution is unknown and different from the training distribution. There are several methods for improving out-of-distribution generalization. According to a survey on the topic, existing methods can be categorized into three parts based on their positions in the whole learning pipeline: unsupervised representation learning, supervised model learning and optimization. Another approach to out-of-distribution generalization is via learning domain-invariant features or hypothesis-invariant features.
 
Out-of-Distribution (OOD) generalization refers to the ability of a machine learning model to generalize to new data that comes from a different distribution than the training data. This is a challenging problem because the testing distribution is unknown and different from the training distribution. There are several methods for improving out-of-distribution generalization. According to a survey on the topic, existing methods can be categorized into three parts based on their positions in the whole learning pipeline: unsupervised representation learning, supervised model learning and optimization. Another approach to out-of-distribution generalization is via learning domain-invariant features or hypothesis-invariant features.
 +
  
 
<youtube>Ugxj_6_Nzug</youtube>
 
<youtube>Ugxj_6_Nzug</youtube>

Revision as of 17:03, 27 May 2023

YouTube ... Quora ...Google search ...Google News ...Bing News

Out-of-Distribution (OOD) generalization refers to the ability of a machine learning model to generalize to new data that comes from a different distribution than the training data. This is a challenging problem because the testing distribution is unknown and different from the training distribution. There are several methods for improving out-of-distribution generalization. According to a survey on the topic, existing methods can be categorized into three parts based on their positions in the whole learning pipeline: unsupervised representation learning, supervised model learning and optimization. Another approach to out-of-distribution generalization is via learning domain-invariant features or hypothesis-invariant features.