Introduction
Core Concept
After we computed the scores, we transform them into probability to give us better intuition about the classifier’s confidence. Next, we
Explanation
Let the be the output of score function, i.e.,
We convert the score into probability using “Softmax Function” The conditional probability represent the score image get on label
Then, the loss of image is defined as
Putting it together we get the equation below:
Problems & Solutions for Cross-Entropy Loss
Q1: What is min/max possible loss A1: min: , max:
Q2: If all scores are small random values, what is the loss for any input? A2: , where is number of labels
At the start of the training, the scores will be small random value, we can then compare the loss value with to check if there's bug in our code