| ID | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| -6.49 | -2.64 | -2.21 | -1.05 | -0.42 | 0.00 | 0.73 | 1.59 | 2.06 | 2.79 | 3.79 | 4.53 | 5.17 | 5.65 | 7.33 | |
| True | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 |
| Pred | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.07 | 0.16 | 0.21 | 0.28 | 0.38 | 0.45 | 0.52 | 0.56 | 0.73 |
| Correct | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✗ | ✗ | ✗ | ✗ | ✗ | ✓ | ✓ | ✓ |
| Confidence | 50.0% | 50.0% | 50.0% | 50.0% | 50.0% | 50.0% | 42.7% | 34.1% | 29.4% | 22.1% | 12.1% | 4.7% | 1.7% | 6.5% | 23.3% |
The Sigmoid (Logistic) Function maps any real value to a probability between 0 and 1:
The sigmoid has several crucial properties: its range is bounded to
Logistic Regression Model: The model computes a weighted combination of inputs and passes it through the sigmoid to obtain a probability:
To classify, we use a threshold (typically 0.5): predict class 1 if
Loss Functions: Unlike regression, we cannot use mean squared error (MSE) for classification because combining MSE with the sigmoid output creates a non-convex cost surface with multiple local minima. Instead, we use cross-entropy loss (also called log loss or negative log-likelihood), which is designed for probabilistic classification and remains convex:
Cross-entropy heavily penalizes confident wrong predictions: predicting probability 0.9 for class 1 when the true label is 0 incurs much larger loss than predicting 0.6. This encourages the model to be calibrated and honest about uncertainty.
Practical guidance: Use sigmoid activation for classification to obtain probabilistic outputs and smooth decision boundaries. The sigmoid ensures predictions stay in the valid probability range [0,1] and provides confidence estimates. Cross-entropy loss is the standard choice for training classification models because it properly handles probabilistic predictions and enables efficient gradient-based optimization.
Developed by Kevin Yu & Panagiotis Angeloudis