1️⃣ Why ML Fails in Production
Most ML models don’t fail loudly.
They slowly rot.
Common reasons:
-
Real-world data changes
-
User behavior evolves
-
Labels arrive late (or never)
-
Monitoring focuses on infra, not predictions
💡 In production, “model trained successfully” ≠ “model still works.”
2️⃣ Data Drift
What it is
Input data changes, but the relationship between input → output stays the same.
📌 Example:
-
Spam model trained on:
-
“FREE”, “WIN”, “CLICK”
-
-
New emails now include:
-
Emojis, short links, slang, QR codes
-
Same concept of spam — different data distribution.
How it looks
-
Feature means shift
-
New categorical values appear
-
Missing values increase
How to detect (automated)
-
Statistical comparison: training vs live data
-
Alert when drift exceeds threshold
Example: Simple Python Drift Check






No comments:
Post a Comment