Loss functions in Deep Learning using Keras

Loss functions in Deep Learning using Keras

In Deep Learning, neural networks requires an optimizer and a loss function to configure an efficient model. The purpose of loss functions is to compute the quantity that a model should seek to minimize during training. Loss function are also termed as Cost function.

Loss functions are categorized  namely, Probabilistic losses and Regression losses.

For training the neural network various algorithm are used. To achieve optimization the weights are updated using back propagation and the optimization algorithms are used to reduce errors in the next iteration with weights changed.

The score calculated after each evaluation is called the loss

Probabilistic  Loss:

These loss functions are used to identify classification based models

Majority used loss functions in this category are


Binary Cross Entropy:

This function calculates the loss of classification model where the target variable is binary like 0 and 1.

 

Categorical Cross Entropy:

This function is used to calculate the loss of classification model where the target variable values are more than two.

 

Sparse Categorical Cross Entropy:

This is similar to Categorical Cross entry but The main difference is the Categorical one has the output in the form of one hot encoded vectors whereas the Sparse Categorical has it in integers.

 

Regression losses:

These loss functions are useful to model the linear relationship between several independent and a dependent variable

 

Mean Square Error:

Mean Square Error(MSE) is the most commonly used regression loss function. MSE is the sum of squared distances between our target variable and predicted values. The MSE loss (Y-axis) reaches its minimum value at prediction (X-axis) = 100. The range is 0 to ∞.

 

Mean Absolute Error:

Mean Absolute Error(MAE) is the sum of absolute differences between our target and predicted variables. So it measures the average magnitude of errors in a set of predictions, without considering their directions. The range is also 0 to ∞.

 

Mean Absolute Percentage Error:

Mean Absolute Percentage Error(MAPE) is also known as mean absolute percentage deviation (MAPD), is a measure of prediction accuracy of a forecasting method in statistics, for example in trend estimation

 

Mean Squared Logarithmic Error:

MSE is a measure of the ratio between the true and predicted values.

Mean squared logarithmic error is a variation of the Mean Squared Error and it only cares about the percentual difference, that means MSLE will treat small fluctuations between small true and predicted value as the same as a big difference between large true and predicted values.

 

Cosine Similarity:

Cosine similarity is a measure of similarity between two non-zero vectors of an inner product space. This loss function Computes the cosine similarity between labels and predictions.

It is just a number between -1 and 1. 

When it is a negative number between -1 and 0, then

0 indicates orthogonality, 

and values closer to -1 indicate greater similarity. 

 

Comments

Popular posts from this blog

Exploring Gini Index in Decision Tree

Optimization techniques in Deep Learning