// Build end-to-end MLOps pipelines from data preparation through model training, validation, and production deployment. Use when creating ML pipelines, implementing MLOps practices, or automating model training and deployment workflows.
| name | ml-pipeline-workflow |
| description | Build end-to-end MLOps pipelines from data preparation through model training, validation, and production deployment. Use when creating ML pipelines, implementing MLOps practices, or automating model training and deployment workflows. |
Complete end-to-end MLOps pipeline orchestration from data preparation through model deployment.
This skill provides comprehensive guidance for building production ML pipelines that handle the full lifecycle: data ingestion → preparation → training → validation → deployment → monitoring.
Pipeline Architecture
Data Preparation
Model Training
Model Validation
Deployment Automation
See the references/ directory for detailed guides:
The assets/ directory contains:
# 1. Define pipeline stages
stages = [
"data_ingestion",
"data_validation",
"feature_engineering",
"model_training",
"model_validation",
"model_deployment"
]
# 2. Configure dependencies
# See assets/pipeline-dag.yaml.template for full example
Data Preparation Phase
Training Phase
Validation Phase
Deployment Phase
Start with the basics and gradually add complexity:
# See assets/pipeline-dag.yaml.template
stages:
- name: data_preparation
dependencies: []
- name: model_training
dependencies: [data_preparation]
- name: model_evaluation
dependencies: [model_training]
- name: model_deployment
dependencies: [model_evaluation]
# Stream processing for real-time features
# Combined with batch training
# See references/data-preparation.md
# Automated retraining on schedule
# Triggered by data drift detection
# See references/model-training.md
After setting up your pipeline: