Difference between revisions of "Backpropagation"

From
Jump to: navigation, search
m
 
(10 intermediate revisions by the same user not shown)
Line 2: Line 2:
 
|title=PRIMO.ai
 
|title=PRIMO.ai
 
|titlemode=append
 
|titlemode=append
|keywords=artificial, intelligence, machine, learning, models, algorithms, data, singularity, moonshot, Tensorflow, Google, Nvidia, Microsoft, Azure, Amazon, AWS  
+
|keywords=ChatGPT, artificial, intelligence, machine, learning, GPT-4, GPT-5, NLP, NLG, NLC, NLU, models, data, singularity, moonshot, Sentience, AGI, Emergence, Moonshot, Explainable, TensorFlow, Google, Nvidia, Microsoft, Azure, Amazon, AWS, Hugging Face, OpenAI, Tensorflow, OpenAI, Google, Nvidia, Microsoft, Azure, Amazon, AWS, Meta, LLM, metaverse, assistants, agents, digital twin, IoT, Transhumanism, Immersive Reality, Generative AI, Conversational AI, Perplexity, Bing, You, Bard, Ernie, prompt Engineering LangChain, Video/Image, Vision, End-to-End Speech, Synthesize Speech, Speech Recognition, Stanford, MIT |description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools
|description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools  
+
 
 +
<!-- Google tag (gtag.js) -->
 +
<script async src="https://www.googletagmanager.com/gtag/js?id=G-4GCWLBVJ7T"></script>
 +
<script>
 +
  window.dataLayer = window.dataLayer || [];
 +
  function gtag(){dataLayer.push(arguments);}
 +
  gtag('js', new Date());
 +
 
 +
  gtag('config', 'G-4GCWLBVJ7T');
 +
</script>
 
}}
 
}}
[http://www.youtube.com/results?search_query=backpropagation Youtube search...]
+
[https://www.youtube.com/results?search_query=backpropagation Youtube search...]
[http://www.google.com/search?q=Backpropagation+deep+machine+learning+ML ...Google search]
+
[https://www.google.com/search?q=Backpropagation+deep+machine+learning+ML ...Google search]
  
* [[Gradient Descent Optimization & Challenges]]
+
* [[Backpropagation]] ... [[Feed Forward Neural Network (FF or FFNN)|FFNN]] ... [[Forward-Forward]] ... [[Activation Functions]] ...[[Softmax]] ... [[Loss]] ... [[Boosting]] ... [[Gradient Descent Optimization & Challenges|Gradient Descent]] ... [[Algorithm Administration#Hyperparameter|Hyperparameter]] ... [[Manifold Hypothesis]] ... [[Principal Component Analysis (PCA)|PCA]]
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
 
* [[Objective vs. Cost vs. Loss vs. Error Function]]
* [http://en.wikipedia.org/wiki/Backpropagation Wikipedia]
+
* [[Optimization Methods]]
* [http://neuralnetworksanddeeplearning.com/chap2.html How the backpropagation algorithm works]
+
* [https://en.wikipedia.org/wiki/Backpropagation Wikipedia]
* [http://hmkcode.github.io/ai/backpropagation-step-by-step/ Backpropagation Step by Step]
+
* [https://neuralnetworksanddeeplearning.com/chap2.html How the backpropagation algorithm works]
* [http://www.unite.ai/what-is-backpropagation/ What is Backpropagation? | Daniel Nelson - Unite.ai]
+
* [https://hmkcode.github.io/ai/backpropagation-step-by-step/ Backpropagation Step by Step]
 +
* [https://www.unite.ai/what-is-backpropagation/ What is Backpropagation? | Daniel Nelson - Unite.ai]
 
* [[Other Challenges]] in Artificial Intelligence
 
* [[Other Challenges]] in Artificial Intelligence
* [http://pathmind.com/wiki/backpropagation A Beginner's Guide to Backpropagation in Neural Networks | Chris Nicholson - A.I. Wiki pathmind]
+
* [https://pathmind.com/wiki/backpropagation A Beginner's Guide to Backpropagation in Neural Networks | Chris Nicholson - A.I. Wiki pathmind]
 
 
  
[http://developers.google.com/machine-learning/glossary/ The primary algorithm for performing gradient descent on neural networks. First, the output values of each node are calculated (and cached) in a forward pass. Then, the partial derivative of the error with respect to each parameter is calculated in a backward pass through the graph. Machine Learning Glossary | Google]
 
  
 +
The primary algorithm for performing gradient descent on neural networks. First, the output values of each node are calculated (and cached) in a forward pass. Then, the partial derivative of the error with respect to each parameter is calculated in a backward pass through the graph. [https://developers.google.com/machine-learning/glossary/ Machine Learning Glossary | Google]
  
http://hmkcode.github.io/images/ai/backpropagation.png
 
  
 +
https://hmkcode.github.io/images/ai/backpropagation.png
  
<youtube>q555kfIFUCM</youtube>
 
<youtube>FaHHWdsIYQg</youtube>
 
 
<youtube>Ilg3gGewQ5U</youtube>
 
<youtube>Ilg3gGewQ5U</youtube>
 
<youtube>An5z8lR8asY</youtube>
 
<youtube>An5z8lR8asY</youtube>
Line 33: Line 40:
 
<youtube>WZDMNM36PsM</youtube>
 
<youtube>WZDMNM36PsM</youtube>
 
<youtube>g9V-MHxSCcs</youtube>
 
<youtube>g9V-MHxSCcs</youtube>
 +
<youtube>q555kfIFUCM</youtube>
 +
<youtube>FaHHWdsIYQg</youtube>

Latest revision as of 09:30, 6 August 2023

Youtube search... ...Google search


The primary algorithm for performing gradient descent on neural networks. First, the output values of each node are calculated (and cached) in a forward pass. Then, the partial derivative of the error with respect to each parameter is calculated in a backward pass through the graph. Machine Learning Glossary | Google


backpropagation.png