Difference between revisions of "Regularization"
| Line 1: | Line 1: | ||
| + | {{#seo: | ||
| + | |title=PRIMO.ai | ||
| + | |titlemode=append | ||
| + | |keywords=artificial, intelligence, machine, learning, models, algorithms, data, singularity, moonshot, Tensorflow, Google, Nvidia, Microsoft, Azure, Amazon, AWS | ||
| + | |description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools | ||
| + | }} | ||
[http://www.youtube.com/results?search_query=Regularization+Overfitting Youtube search...] | [http://www.youtube.com/results?search_query=Regularization+Overfitting Youtube search...] | ||
[http://www.google.com/search?q=Regularization+deep+machine+learning+ML ...Google search] | [http://www.google.com/search?q=Regularization+deep+machine+learning+ML ...Google search] | ||
Revision as of 13:24, 3 February 2019
Youtube search... ...Google search
Good practices for addressing the Overfitting Challenge:
- add more data
- use Data Augmentation
- use Batch Norm(alization) & Standardization
- use architectures that generalize well
- reduce architecture complexity
- add Regularization
- L1 and L2 Regularization - update the general cost function by adding another term known as the regularization term.
- Dropout - at every iteration, it randomly selects some nodes and temporarily removes the nodes (along with all of their incoming and outgoing connections)
- Data Augmentation
- Early Stopping
Regularization is a technique which makes slight modifications to the learning algorithm such that the model generalizes better. This in turn improves the model’s performance on the unseen data as well. An Overview of Regularization Techniques in Deep Learning (with Python code) | Shubham Jain