Difference between revisions of "Backpropagation"

From
Jump to: navigation, search
m
 
(20 intermediate revisions by the same user not shown)
Line 1: Line 1:
[http://www.youtube.com/results?search_query=backpropagation Youtube search...]
+
{{#seo:
 +
|title=PRIMO.ai
 +
|titlemode=append
 +
|keywords=ChatGPT, artificial, intelligence, machine, learning, GPT-4, GPT-5, NLP, NLG, NLC, NLU, models, data, singularity, moonshot, Sentience, AGI, Emergence, Moonshot, Explainable, TensorFlow, Google, Nvidia, Microsoft, Azure, Amazon, AWS, Hugging Face, OpenAI, Tensorflow, OpenAI, Google, Nvidia, Microsoft, Azure, Amazon, AWS, Meta, LLM, metaverse, assistants, agents, digital twin, IoT, Transhumanism, Immersive Reality, Generative AI, Conversational AI, Perplexity, Bing, You, Bard, Ernie, prompt Engineering LangChain, Video/Image, Vision, End-to-End Speech, Synthesize Speech, Speech Recognition, Stanford, MIT |description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools 
  
* [[Gradient Descent Optimization & Challenges]]
+
<!-- Google tag (gtag.js) -->
 +
<script async src="https://www.googletagmanager.com/gtag/js?id=G-4GCWLBVJ7T"></script>
 +
<script>
 +
  window.dataLayer = window.dataLayer || [];
 +
  function gtag(){dataLayer.push(arguments);}
 +
  gtag('js', new Date());
 +
 
 +
  gtag('config', 'G-4GCWLBVJ7T');
 +
</script>
 +
}}
 +
[https://www.youtube.com/results?search_query=backpropagation Youtube search...]
 +
[https://www.google.com/search?q=Backpropagation+deep+machine+learning+ML ...Google search]
 +
 
 +
* [[Backpropagation]] ... [[Feed Forward Neural Network (FF or FFNN)|FFNN]] ... [[Forward-Forward]] ... [[Activation Functions]] ...[[Softmax]] ... [[Loss]] ... [[Boosting]] ... [[Gradient Descent Optimization & Challenges|Gradient Descent]] ... [[Algorithm Administration#Hyperparameter|Hyperparameter]] ... [[Manifold Hypothesis]] ... [[Principal Component Analysis (PCA)|PCA]]
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
 +
* [[Optimization Methods]]
 +
* [https://en.wikipedia.org/wiki/Backpropagation Wikipedia]
 +
* [https://neuralnetworksanddeeplearning.com/chap2.html How the backpropagation algorithm works]
 +
* [https://hmkcode.github.io/ai/backpropagation-step-by-step/ Backpropagation Step by Step]
 +
* [https://www.unite.ai/what-is-backpropagation/ What is Backpropagation? | Daniel Nelson - Unite.ai]
 +
* [[Other Challenges]] in Artificial Intelligence
 +
* [https://pathmind.com/wiki/backpropagation A Beginner's Guide to Backpropagation in Neural Networks | Chris Nicholson - A.I. Wiki pathmind]
 +
 +
 +
The primary algorithm for performing gradient descent on neural networks. First, the output values of each node are calculated (and cached) in a forward pass. Then, the partial derivative of the error with respect to each parameter is calculated in a backward pass through the graph. [https://developers.google.com/machine-learning/glossary/ Machine Learning Glossary | Google]
 +
 +
 +
https://hmkcode.github.io/images/ai/backpropagation.png
  
<youtube>q555kfIFUCM</youtube>
 
<youtube>FaHHWdsIYQg</youtube>
 
 
<youtube>Ilg3gGewQ5U</youtube>
 
<youtube>Ilg3gGewQ5U</youtube>
 
<youtube>An5z8lR8asY</youtube>
 
<youtube>An5z8lR8asY</youtube>
Line 12: Line 39:
 
<youtube>XE3krf3CQls</youtube>
 
<youtube>XE3krf3CQls</youtube>
 
<youtube>WZDMNM36PsM</youtube>
 
<youtube>WZDMNM36PsM</youtube>
 +
<youtube>g9V-MHxSCcs</youtube>
 +
<youtube>q555kfIFUCM</youtube>
 +
<youtube>FaHHWdsIYQg</youtube>

Latest revision as of 09:30, 6 August 2023

Youtube search... ...Google search


The primary algorithm for performing gradient descent on neural networks. First, the output values of each node are calculated (and cached) in a forward pass. Then, the partial derivative of the error with respect to each parameter is calculated in a backward pass through the graph. Machine Learning Glossary | Google


backpropagation.png