Introduction

Core Concept

After we computed the scores, we transform them into probability to give us better intuition about the classifier’s confidence. Next, we

Explanation

Let the be the output of score function, i.e.,

We convert the score into probability using “Softmax Function” The conditional probability represent the score image get on label

Then, the loss of image is defined as

Putting it together we get the equation below:


Problems & Solutions for Cross-Entropy Loss

Q1: What is min/max possible loss A1: min: , max:

Q2: If all scores are small random values, what is the loss for any input? A2: , where is number of labels

At the start of the training, the scores will be small random value, we can then compare the loss value with to check if there's bug in our code