VisionWatch runs a live dashboard that shows safety detections from warehouse cameras. A binary object detection model scores each candidate event from 0 to 1, and the dashboard only displays detections above a chosen threshold. Operations managers complain that the current setting creates too many distracting alerts, while safety leads worry that important events are being hidden.
Validation set: 12,000 candidate events, 600 true incidents (5% prevalence).
| Threshold | Precision | Recall | F1 | False Positives/day | True Positives/day | Alerts/day |
|---|---|---|---|---|---|---|
| 0.30 | 0.41 | 0.93 | 0.57 | 420 | 295 | 715 |
| 0.50 | 0.68 | 0.81 | 0.74 | 115 | 257 | 372 |
| 0.70 | 0.86 | 0.54 | 0.66 | 32 | 171 | 203 |
| 0.85 | 0.94 | 0.31 | 0.47 | 9 | 98 | 107 |
Additional model metrics on the same validation set:
| Metric | Value |
|---|---|
| AUC-ROC | 0.91 |
| PR-AUC | 0.78 |
| Brier Score | 0.16 |
| Expected Calibration Error | 0.09 |
You need to recommend the right threshold for a live dashboard where users can only review about 250 alerts per day. Missing a true safety incident is estimated to cost 20x more than showing a false alert, but repeated false alerts reduce trust and dashboard usage.