Difference between revisions of "Backpropagation"

From
Jump to: navigation, search
Line 10: Line 10:
 
* [[Gradient Descent Optimization & Challenges]]
 
* [[Gradient Descent Optimization & Challenges]]
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
 +
* [http://en.wikipedia.org/wiki/Backpropagation Wikipedia]
  
 
[http://developers.google.com/machine-learning/glossary/ The primary algorithm for performing gradient descent on neural networks. First, the output values of each node are calculated (and cached) in a forward pass. Then, the partial derivative of the error with respect to each parameter is calculated in a backward pass through the graph. Machine Learning Glossary | Google]
 
[http://developers.google.com/machine-learning/glossary/ The primary algorithm for performing gradient descent on neural networks. First, the output values of each node are calculated (and cached) in a forward pass. Then, the partial derivative of the error with respect to each parameter is calculated in a backward pass through the graph. Machine Learning Glossary | Google]

Revision as of 21:58, 11 February 2019