Difference between revisions of "Overfitting Challenge"

From
Jump to: navigation, search
m (BPeat moved page Regularization to Regularization - The Problem Of Overfitting without leaving a redirect)
m
 
(38 intermediate revisions by the same user not shown)
Line 1: Line 1:
[http://www.youtube.com/results?search_query=Gradient+Boosting+Algorithms Youtube search...]
+
{{#seo:
 +
|title=PRIMO.ai
 +
|titlemode=append
 +
|keywords=artificial, intelligence, machine, learning, models, algorithms, data, singularity, moonshot, Tensorflow, Google, Nvidia, Microsoft, Azure, Amazon, AWS
 +
|description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools
 +
}}
 +
[http://www.youtube.com/results?search_query=Regularization+Dropout+Overfitting Youtube search...]
 +
[http://www.google.com/search?q=Regularization+Dropout+deep+machine+learning+ML ...Google search]
  
* [[Gradient Descent Optimization & Challenges]]
+
* [http://www.unite.ai/what-is-overfitting/ What Is Overfitting? | Daniel Nelson - Unite.ai]
* [[Objective vs. Cost vs. Loss vs. Error Function]]
+
* [http://docs.aws.amazon.com/machine-learning/latest/dg/evaluating_models.html#overfitting Preventing Overfitting | AWS]
 +
* [[Bias and Variances]]
  
Gradient Boosting Algorithm uses multiple weak algorithms to create a more powerful accurate algorithm. Instead of using a single estimator, having multiple will create a more stable and robust algorithm. The specialty of Gradient Boosting Algorithms is their higher accuracy. There are several Gradient Boosting Algorithms. [http://towardsdatascience.com/10-machine-learning-algorithms-you-need-to-know-77fb0055fe0 10 Machine Learning Algorithms You need to Know | Sidath Asir @ Medium]
+
Overfitting is a problem in machine learning in general, not just in neural networks. The problem is inherent in the way machine learning models are developed: A set of "training data" is used to "train" the model. The goal is to have a model that can then be used on data that hasn't been seen before. Over-fitting refers to the problem of having the model trained to work so well on the training data that it starts to work more poorly on data it hasn't seen before. There are a number of techniques to mitigate or prevent over-fitting. [http://wiki.fast.ai/index.php/Over-fitting | Deep Learning Course Wiki]
  
* [http://xgboost.readthedocs.io/en/latest/ XGBoost] — uses liner and tree algorithms
+
* [[Approach to Bias and Variances|Bias]] is the simplifying assumptions made by the model to make the target function easier to approximate.  
* [http://lightgbm.readthedocs.io/en/latest/ LightGBM] — uses only tree-based algorithms; has incredible high performance as well.
+
* Variance is the amount that the estimate of the target function will change given different training data.  
 +
Trade-off is tension between the error introduced by the bias and the variance. [http://machinelearningmastery.com/gentle-introduction-to-the-bias-variance-trade-off-in-machine-learning/ Jason Brownlee]
  
<youtube>sRktKszFmSk</youtube>
+
http://miro.medium.com/max/468/1*xwtSpR_zg7j7zusa4IDHNQ.png
 +
 
 +
Whenever we discuss model prediction, it’s important to understand prediction errors (bias and variance). There is a tradeoff between a model’s ability to minimize bias and variance... To build a good model, we need to find a good balance between bias and variance such that it minimizes the total error.[http://towardsdatascience.com/understanding-the-bias-variance-tradeoff-165e6942b229 Understanding the Bias-Variance Tradeoff | Seema Singh]
 +
 
 +
Good practices for addressing overfitting:
 +
 
 +
* add more data
 +
* use [[Data Quality#Batch Norm(alization) & Standardization|Batch Norm(alization) & Standardization]]
 +
* use architectures that generalize well
 +
* reduce architecture complexity
 +
* add [[Regularization]]
 +
** [[L1 and L2 Regularization]] -  update the general cost function by adding another term known as the regularization term.
 +
** [[Dropout]] - at every iteration, it randomly selects some nodes and temporarily removes the nodes (along with all of their incoming and outgoing connections)
 +
** [[Data Quality#Data Augmentation, Data Labeling, and Auto-Tagging|Data Augmentation]]
 +
** [[Early Stopping]]
 +
 
 +
http://s3-ap-south-1.amazonaws.com/av-blog-media/wp-content/uploads/2018/04/Screen-Shot-2018-04-03-at-7.52.01-PM-e1522832332857.png
 +
http://s3-ap-south-1.amazonaws.com/av-blog-media/wp-content/uploads/2018/04/Screen-Shot-2018-04-04-at-2.43.37-PM-768x592.png
 +
 
 +
<youtube>cJA5IHIIL30</youtube>
 +
<youtube>ms-Ooh9mjiE</youtube>
 +
<youtube>-JopeGg60QY</youtube>
 +
<youtube>S4ZUwgesjS8</youtube>
 +
<youtube>g2ext963IZg</youtube>
 +
<youtube>EuBBz3bI-aA</youtube>
 +
<youtube>Q81RR3yKn30</youtube>

Latest revision as of 19:13, 19 September 2020

Youtube search... ...Google search

Overfitting is a problem in machine learning in general, not just in neural networks. The problem is inherent in the way machine learning models are developed: A set of "training data" is used to "train" the model. The goal is to have a model that can then be used on data that hasn't been seen before. Over-fitting refers to the problem of having the model trained to work so well on the training data that it starts to work more poorly on data it hasn't seen before. There are a number of techniques to mitigate or prevent over-fitting. | Deep Learning Course Wiki

  • Bias is the simplifying assumptions made by the model to make the target function easier to approximate.
  • Variance is the amount that the estimate of the target function will change given different training data.

Trade-off is tension between the error introduced by the bias and the variance. Jason Brownlee

1*xwtSpR_zg7j7zusa4IDHNQ.png

Whenever we discuss model prediction, it’s important to understand prediction errors (bias and variance). There is a tradeoff between a model’s ability to minimize bias and variance... To build a good model, we need to find a good balance between bias and variance such that it minimizes the total error.Understanding the Bias-Variance Tradeoff | Seema Singh

Good practices for addressing overfitting:

Screen-Shot-2018-04-03-at-7.52.01-PM-e1522832332857.png Screen-Shot-2018-04-04-at-2.43.37-PM-768x592.png