Difference between revisions of "Out-of-Distribution (OOD) Generalization"

From
Jump to: navigation, search
m
m
Line 25: Line 25:
  
  
 +
<youtube>W3XE9yD5H4A</youtube>
 
<youtube>Ugxj_6_Nzug</youtube>
 
<youtube>Ugxj_6_Nzug</youtube>
 
<youtube>CxUmPZMg858</youtube>
 
<youtube>CxUmPZMg858</youtube>
Line 32: Line 33:
 
<youtube>v27iTSkYugU</youtube>
 
<youtube>v27iTSkYugU</youtube>
 
<youtube>_ZpBgkpgPp8</youtube>
 
<youtube>_ZpBgkpgPp8</youtube>
<youtube>W3XE9yD5H4A</youtube>
 

Revision as of 17:10, 27 May 2023

YouTube ... Quora ...Google search ...Google News ...Bing News

Out-of-Distribution (OOD) generalization refers to the ability of a machine learning model to generalize to new data that comes from a different distribution than the training data. This is a challenging problem because the testing distribution is unknown and different from the training distribution. There are several methods for improving out-of-distribution generalization. According to a survey on the topic, existing methods can be categorized into three parts based on their positions in the whole learning pipeline: unsupervised representation learning, supervised model learning and optimization. Another approach to out-of-distribution generalization is via learning domain-invariant features or hypothesis-invariant features.