StreamBox uses a binary classification model to predict which subscribers are likely to cancel in the next 30 days. Customers above a score threshold receive a retention offer worth $8, but unnecessary offers reduce margin, so the team wants to tune the decision threshold rather than retrain the model.
Validation set size: 20,000 users. Positive class = churn. Current threshold = 0.50.
| Metric | Threshold 0.30 | Threshold 0.50 | Threshold 0.70 |
|---|---|---|---|
| Precision | 0.41 | 0.62 | 0.79 |
| Recall | 0.86 | 0.58 | 0.31 |
| F1 Score | 0.56 | 0.60 | 0.45 |
| Accuracy | 0.71 | 0.84 | 0.89 |
| Users flagged for offer | 4,200 | 2,100 | 950 |
| True churners caught | 1,720 | 1,160 | 620 |
| False positives | 2,480 | 940 | 330 |
The marketing lead is asking why the team should change the threshold if the model itself has not changed. They also want to know which threshold best balances churn prevention against offer cost.