Detection Strategies
Concept drift is harder to detect because it requires ground truth labels, which are often delayed (fraud labels arrive weeks later, medical outcomes take months). Strategies: Direct monitoring — when labels are available, track model accuracy, precision, recall, and F1 over time. A sustained drop signals concept drift. Prediction drift — when labels are unavailable, monitor the distribution of model predictions. If the model suddenly predicts “fraud” 3x more often, something changed. Proxy metrics — use business metrics as proxies (conversion rate, customer complaints, escalation rate). Window comparison — compare model performance on recent data vs. older data using a sliding window.
Concept Drift Types
// Types of concept drift
Sudden drift:
┌──────┐
│ ████ │████████
│ ████ │ ████████
└──────┘
// Abrupt change (e.g., new regulation)
Gradual drift:
████████
████████
████████
████████
// Slow transition (e.g., user behavior)
Recurring drift:
████ ████ ████
████ ████
// Seasonal patterns (e.g., holiday sales)
Detection without labels:
Monitor prediction distribution
Track confidence scores over time
Use business metrics as proxies
Compare recent vs. baseline windows
Key insight: Recurring drift (seasonal patterns) is often mistaken for concept drift. Before retraining, check if the pattern is seasonal. If your model sees this pattern every December, it’s not drift — it’s a known cycle that should be modeled explicitly.