Difference between revisions of "Backpropagation"

From
Jump to: navigation, search
m
m
Line 8: Line 8:
 
[http://www.google.com/search?q=Backpropagation+deep+machine+learning+ML ...Google search]
 
[http://www.google.com/search?q=Backpropagation+deep+machine+learning+ML ...Google search]
  
* [[Backpropagation]]   ... [[Feed Forward Neural Network (FF or FFNN)]]   ... [[Forward-Forward]]
+
* [[Backpropagation]] ...[[Gradient Descent Optimization & Challenges]]  ...[[Feed Forward Neural Network (FF or FFNN)]] ...[[Forward-Forward]]
* [[Gradient Descent Optimization & Challenges]]
 
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
 
* [http://en.wikipedia.org/wiki/Backpropagation Wikipedia]
 
* [http://en.wikipedia.org/wiki/Backpropagation Wikipedia]

Revision as of 12:59, 26 February 2023

Youtube search... ...Google search


The primary algorithm for performing gradient descent on neural networks. First, the output values of each node are calculated (and cached) in a forward pass. Then, the partial derivative of the error with respect to each parameter is calculated in a backward pass through the graph. Machine Learning Glossary | Google


backpropagation.png