Binary cross-entropy loss
Revision as of 15:36, 3 February 2023 by KevinYager (talk | contribs) (Created page with "The binary cross-entropy loss is given by: :<math> L = - \frac{1}{m} \sum_{i=1}^{m} \left[ y_i \cdot \log{ (\hat{y_i}) } + (1-y_i) \cdot \log{ (1-\hat{y_i}) } \right] </math>...")
The binary cross-entropy loss is given by:
for training examples (indexed by ) where is the class label (0 or 1) and is the prediction for that example (i.e. the predicted probability that it is a positive example). Thus is the probability that it is a negative example.