Difference between revisions of "Gradient Descent Optimization & Challenges"

From
Jump to: navigation, search
Line 13: Line 13:
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
 
* [[Topology and Weight Evolving Artificial Neural Network (TWEANN)]]
 
* [[Topology and Weight Evolving Artificial Neural Network (TWEANN)]]
 +
 +
 +
http://cdn-images-1.medium.com/max/800/1*NRCWfdXa7b-ak2nBtmwRvw.png
  
 
== Gradient Descent - Stochastic (SGD), Batch (BGD) & Mini-Batch ==
 
== Gradient Descent - Stochastic (SGD), Batch (BGD) & Mini-Batch ==

Revision as of 22:09, 16 February 2019

YouTube search... ...Google search


1*NRCWfdXa7b-ak2nBtmwRvw.png

Gradient Descent - Stochastic (SGD), Batch (BGD) & Mini-Batch

Youtube search...

Vanishing & Exploding Gradients Problems

YouTube Search

Vanishing & Exploding Gradients Challenges with Long Short-Term Memory (LSTM) and Recurrent Neural Networks (RNN)

[YouTube Search]