MLOps Interview Questions
30+ MLOps interview questions with hidden answers, organized by topic. Click "Show Answer" to reveal. Perfect for a quick 15-minute revision before an ML engineering interview.
MLOps Fundamentals
Q: What is MLOps and how does it differ from DevOps?
Q: Describe Google's three MLOps maturity levels.
Q: Why do most ML projects fail to reach production?
Q: What are the key differences between ML systems and traditional software systems?
Q: What is technical debt specific to ML systems?
Want deeper coverage? See MLOps Overview and Core Concepts.
Experiment Tracking & Versioning
Q: Why is experiment tracking important and what should you log?
Q: How do you version datasets that are too large for Git?
.dvc) in Git while the actual data lives in cloud storage (S3, GCS, Azure Blob). dvc add data/ creates the pointer, git commit versions it, dvc push uploads the data. git checkout v2.0 && dvc checkout restores the exact dataset for any historical version. Alternative: LakeFS provides Git-like branching for data lakes.Q: What is the difference between a model registry and experiment tracking?
Q: How do you ensure reproducibility in ML experiments?
Pipelines & Feature Stores
Q: What are the stages of a typical ML training pipeline?
Q: What is training/serving skew and how do you prevent it?
Q: Compare Kubeflow Pipelines, Apache Airflow, and SageMaker Pipelines for ML workflows.
Q: What is point-in-time correctness in a feature store?
Q: What is the difference between the online and offline store in a feature store?
Deeper coverage: ML Pipelines & Feature Stores
Model Deployment
Q: What are the main deployment strategies for ML models?
Q: What's the difference between real-time and batch model serving?
Q: How do you do A/B testing with ML models, and what metrics do you measure?
Q: How do you containerize an ML model for deployment?
mlflow models build-docker -m models:/fraud-detector/Production -n fraud-detector:latest.Q: What is model serving latency optimization?
Deeper coverage: ML Model Lifecycle
Model Monitoring
Q: What types of drift should you monitor for production ML models?
Q: How do you monitor a model when ground truth labels are delayed?
Q: What is PSI (Population Stability Index) and how is it used?
Q: What metrics should an ML model monitoring dashboard include?
Scenario Questions
Q: You inherit an ML model running in production with no documentation. How do you assess its health?
Q: Your fraud detection model's precision dropped from 92% to 78% over 3 months. Diagnose and fix it.
Q: Design an MLOps architecture for a recommendation system serving 10M users.
Q: How would you set up CI/CD for an ML project?
Q: Your model works great in the notebook but fails in production. Common causes?
Q: How do you handle model governance in a regulated industry (finance, healthcare)?