Difference between revisions of "Loss"
Line 14: | Line 14: | ||
* [http://ml-cheatsheet.readthedocs.io/en/latest/loss_functions.html Loss Functions | ML Cheatsheet] | * [http://ml-cheatsheet.readthedocs.io/en/latest/loss_functions.html Loss Functions | ML Cheatsheet] | ||
− | There are many options for loss in Tensorflow (Keras). The actual optimized objective is the mean of the output array across all datapoints. | + | There are many options for loss in Tensorflow (Keras). The actual optimized objective is the mean of the output array across all datapoints. A loss function gives a distance between a model's predictions to the ground truth labels. This is the distance (loss value) that the network aims to minimize; the lower this value, the better the current model describes our training data set. [http://keras.io/losses/ Click here For a list of Keras loss functions.] Loss is one of the two parameters required to compile a model... |
---- | ---- |
Revision as of 09:23, 31 August 2019
YouTube search... ...Google search
- Optimizer Functions
- Objective vs. Cost vs. Loss vs. Error Function
- Common Loss functions in machine learning | Ravindra Parmar - Towards data Science
- Loss Functions Explained | Siraj Raval
- Loss Functions | ML Cheatsheet
There are many options for loss in Tensorflow (Keras). The actual optimized objective is the mean of the output array across all datapoints. A loss function gives a distance between a model's predictions to the ground truth labels. This is the distance (loss value) that the network aims to minimize; the lower this value, the better the current model describes our training data set. Click here For a list of Keras loss functions. Loss is one of the two parameters required to compile a model...