Difference between revisions of "Softmax"
m |
m |
||
| Line 2: | Line 2: | ||
|title=PRIMO.ai | |title=PRIMO.ai | ||
|titlemode=append | |titlemode=append | ||
| − | |keywords=artificial, intelligence, machine, learning, models, algorithms, data, singularity, moonshot, TensorFlow, Google, Nvidia, Microsoft, Azure, Amazon, AWS, Facebook | + | |keywords=artificial, intelligence, machine, learning, models, algorithms, data, singularity, moonshot, TensorFlow, Google, Nvidia, Microsoft, Azure, Amazon, AWS, Meta, Facebook |
|description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools | |description=Helpful resources for your journey with artificial intelligence; videos, articles, techniques, courses, profiles, and tools | ||
}} | }} | ||
Revision as of 22:45, 8 February 2023
Youtube search... ...Google search
- Dimensional Reduction
- Pooling / Sub-sampling: Max, Mean
- (Deep) Convolutional Neural Network (DCNN/CNN)
- Activation Functions
- Multi-Class Neural Networks: Softmax
- The Softmax Function, Simplified - How a regression formula improves accuracy of deep learning models | Hamza Mahmood - Towards Data Science
- Difference Between Softmax Function And Sigmoid Function | Saimadhu Polamuri - Dataaspirant
Function is used in various multiclass classification methods, such as multinomial logistic regression (also known as softmax regression), multiclass linear discriminant analysis, Naive Bayes classifiers, and artificial neural networks.