Difference between revisions of "Optimization Methods"
| Line 10: | Line 10: | ||
* [[Natural Language Processing (NLP)]] | * [[Natural Language Processing (NLP)]] | ||
* [[Recurrent Neural Network (RNN)]] | * [[Recurrent Neural Network (RNN)]] | ||
| − | ** [[Average-SGD Weight-Dropped LSTM (AWD-LSTM)]] | + | ** [[Average-Stochastic Gradient Descent (SGD) Weight-Dropped LSTM (AWD-LSTM)]] |
* Gradient [[Boosting]] Algorithms | * Gradient [[Boosting]] Algorithms | ||
Revision as of 14:47, 27 July 2020
Youtube search... ...Google search
- Natural Language Processing (NLP)
- Recurrent Neural Network (RNN)
- Gradient Boosting Algorithms
Methods:
- Stochastic gradient descent (SGD) (with and without momentum)
- L-BGFS
- Adagrad
- Adadelta
- Root Mean Squared (RMSprop)
- Adam
- Hessian-free (HF)