ai-mlops

MLOps & ML Security - Complete Reference (Jan 2026)

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "ai-mlops" with this command: npx skills add vasilyu1983/ai-agents-public/vasilyu1983-ai-agents-public-ai-mlops

MLOps & ML Security - Complete Reference (Jan 2026)

Production ML lifecycle with modern security practices.

This skill covers:

  • Production: Data ingestion, deployment, drift detection, monitoring, incident response

  • Security: Prompt injection, jailbreak defense, RAG security, output filtering

  • Governance: Privacy protection, supply chain security, safety evaluation

  • Data ingestion (dlt): Load data from APIs, databases to warehouses

  • Model deployment: Batch jobs, real-time APIs, hybrid systems, event-driven automation

  • Operations: Real-time monitoring, drift detection, automated retraining, incident response

Modern Best Practices (Jan 2026):

It is execution-focused:

  • Data ingestion patterns (REST APIs, database replication, incremental loading)

  • Deployment patterns (batch, online, hybrid, streaming, event-driven)

  • Automated monitoring with real-time drift detection

  • Automated retraining pipelines (monitor → detect → trigger → validate → deploy)

  • Incident handling with validated rollback and postmortems

  • Links to copy-paste templates in assets/

Quick Reference

Task Tool/Framework Command When to Use

Data Ingestion dlt (data load tool) dlt pipeline run , dlt init

Loading from APIs, databases to warehouses

Batch Deployment Airflow, Dagster, Prefect airflow dags trigger , dagster job launch

Scheduled predictions on large datasets

API Deployment FastAPI, Flask, TorchServe uvicorn app:app , torchserve --start

Real-time inference (<500ms latency)

LLM Serving vLLM, TGI, BentoML vllm serve model , bentoml serve

High-throughput LLM inference

Model Registry MLflow, W&B, ZenML mlflow.register_model() , zenml model register

Versioning and promoting models

Drift Detection Statistical tests + monitors PSI/KS, embedding drift, prediction drift Detect data/process changes and trigger review

Monitoring Prometheus, Grafana prometheus.yml , Grafana dashboards Metrics, alerts, SLO tracking

AgentOps AgentOps, Langfuse, LangSmith agentops.init() , trace visualization AI agent observability, session replay

Incident Response Runbooks, PagerDuty Documented playbooks, alert routing Handling failures and degradation

Use This Skill When

Use this skill when the user asks for deployment, operations, monitoring, incident handling, or governance for ML/LLM/agent systems, e.g.:

  • "How do I deploy this model to prod?"

  • "Design a batch + online scoring architecture."

  • "Add monitoring and drift detection to our model."

  • "Write an incident runbook for this ML service."

  • "Package this LLM/RAG pipeline as an API."

  • "Plan our retraining and promotion workflow."

  • "Load data from Stripe API to Snowflake."

  • "Set up incremental database replication with dlt."

  • "Build an ELT pipeline for warehouse loading."

If the user is asking only about EDA, modelling, or theory, prefer:

  • ai-ml-data-science (EDA, features, modelling, SQL transformation with SQLMesh)

  • ai-llm (prompting, fine-tuning, eval)

  • ai-rag (retrieval pipeline design)

  • ai-llm-inference (compression, spec decode, serving internals)

If the user is asking about SQL transformation (after data is loaded), prefer:

  • ai-ml-data-science (SQLMesh templates for staging, intermediate, marts layers)

Decision Tree: Choosing Deployment Strategy

User needs to deploy: [ML System] ├─ Data Ingestion? │ ├─ From REST APIs? → dlt REST API templates │ ├─ From databases? → dlt database sources (PostgreSQL, MySQL, MongoDB) │ └─ Incremental loading? → dlt incremental patterns (timestamp, ID-based) │ ├─ Model Serving? │ ├─ Latency <500ms? → FastAPI real-time API │ ├─ Batch predictions? → Airflow/Dagster batch pipeline │ └─ Mix of both? → Hybrid (batch features + online scoring) │ ├─ Monitoring & Ops? │ ├─ Drift detection? → Evidently + automated retraining triggers │ ├─ Performance tracking? → Prometheus + Grafana dashboards │ └─ Incident response? → Runbooks + PagerDuty alerts │ └─ LLM/RAG Production? ├─ Cost optimization? → Caching, prompt templates, token budgets └─ Safety? → See ai-mlops skill

Core Concepts (Vendor-Agnostic)

  • Lifecycle loop: train → validate → deploy → monitor → respond → retrain/retire.

  • Risk controls: access control, data minimization, logging, and change management (NIST AI RMF: https://nvlpubs.nist.gov/nistpubs/ai/NIST.AI.100-1.pdf).

  • Observability planes: system metrics (latency/errors), data metrics (freshness/drift), quality metrics (model performance).

  • Incident readiness: detection, containment, rollback, and root-cause analysis.

Do / Avoid

Do

  • Do gate deployments with repeatable checks: evaluation pass, load test, security review, rollback plan.

  • Do version everything: code, data, features, model artifact, prompt templates, configuration.

  • Do define SLOs and budgets (latency/cost/error rate) before optimizing.

Avoid

  • Avoid manual “clickops” deployments without audit trail.

  • Avoid silent upgrades; require eval + canary for model/prompt changes.

  • Avoid drift dashboards without actions; every alert needs an owner and runbook.

Core Patterns Overview

This skill provides production-ready patterns and guides organized into comprehensive references:

Data & Infrastructure Patterns

Pattern 0: Data Contracts, Ingestion & Lineage → See Data Ingestion Patterns

  • Data contracts with SLAs and versioning

  • Ingestion modes (CDC, batch, streaming)

  • Lineage tracking and schema evolution

  • Replay and backfill procedures

Pattern 1: Choose Deployment Mode → See Deployment Patterns

  • Decision table (batch, online, hybrid, streaming)

  • When to use each mode

  • Deployment mode selection checklist

Pattern 2: Standard Deployment Lifecycle → See Deployment Lifecycle

  • Pre-deploy, deploy, observe, operate, evolve phases

  • Environment promotion (dev → staging → prod)

  • Gradual rollout strategies (canary, blue-green)

Pattern 3: Packaging & Model Registry → See Model Registry Patterns

  • Model registry structure and metadata

  • Packaging strategies (Docker, ONNX, MLflow)

  • Promotion flows (experimental → production)

  • Versioning and governance

Serving Patterns

Pattern 4: Batch Scoring Pipeline → See Deployment Patterns

  • Orchestration with Airflow/Dagster

  • Idempotent scoring jobs

  • Validation and backfill procedures

Pattern 5: Real-Time API Scoring → See API Design Patterns

  • Service design (HTTP/JSON, gRPC)

  • Input/output schemas

  • Rate limiting, timeouts, circuit breakers

Pattern 6: Hybrid & Feature Store Integration → See Feature Store Patterns

  • Batch vs online features

  • Feature store architecture

  • Training-serving consistency

  • Point-in-time correctness

Operations Patterns

Pattern 7: Monitoring & Alerting → See Monitoring Best Practices

  • Data, performance, and technical metrics

  • SLO definition and tracking

  • Dashboard design and alerting strategies

Pattern 8: Drift Detection & Automated Retraining → See Drift Detection Guide

  • Automated retraining triggers

  • Event-driven retraining pipelines

Pattern 9: Incidents & Runbooks → See Incident Response Playbooks

  • Common failure modes

  • Detection, diagnosis, resolution

  • Post-mortem procedures

Pattern 10: LLM / RAG in Production → See LLM & RAG Production Patterns

  • Prompt and configuration management

  • Safety and compliance (PII, jailbreaks)

  • Cost optimization (token budgets, caching)

  • Monitoring and fallbacks

Pattern 11: Cross-Region, Residency & Rollback → See Multi-Region Patterns

  • Multi-region deployment architectures

  • Data residency and tenant isolation

  • Disaster recovery and failover

  • Regional rollback procedures

Pattern 12: Online Evaluation & Feedback Loops → See Online Evaluation Patterns

  • Feedback signal collection (implicit, explicit)

  • Shadow and canary deployments

  • A/B testing with statistical significance

  • Human-in-the-loop labeling

  • Automated retraining cadence

Pattern 13: AgentOps (AI Agent Operations) → See AgentOps Patterns

  • Session tracing and replay for AI agents

  • Cost and latency tracking across agent runs

  • Multi-agent visualization and debugging

  • Tool invocation monitoring

  • Integration with CrewAI, LangGraph, OpenAI Agents SDK

Pattern 14: Edge MLOps & TinyML → See Edge MLOps Patterns

  • Device-aware CI/CD pipelines

  • OTA model updates with rollback

  • Federated learning operations

  • Edge drift detection

  • Intermittent connectivity handling

Resources (Detailed Guides)

For comprehensive operational guides, see:

Core Infrastructure:

  • Data Ingestion Patterns - Data contracts, CDC, batch/streaming ingestion, lineage, schema evolution

  • Deployment Lifecycle - Pre-deploy validation, environment promotion, gradual rollout, rollback

  • Model Registry Patterns - Versioning, packaging, promotion workflows, governance

  • Feature Store Patterns - Batch/online features, hybrid architectures, consistency, latency optimization

Serving & APIs:

  • Deployment Patterns - Batch, online, hybrid, streaming deployment strategies and architectures

  • API Design Patterns - ML/LLM/RAG API patterns, input/output schemas, reliability patterns, versioning

Operations & Reliability:

  • Monitoring Best Practices - Metrics collection, alerting strategies, SLO definition, dashboard design

  • Drift Detection Guide - Statistical tests, automated detection, retraining triggers, recovery strategies

  • Incident Response Playbooks - Runbooks for common failure modes, diagnostics, resolution steps

Security & Governance:

  • Threat Models - Trust boundaries, attack surface, control mapping

  • Prompt Injection Mitigation - Input hardening, tool/RAG containment, least privilege

  • Jailbreak Defense - Robust refusal behavior, safe completion patterns

  • RAG Security - Retrieval poisoning, context injection, sensitive data leakage

  • Output Filtering - Layered filters (PII/toxicity/policy), block/rewrite strategies

  • Privacy Protection - PII handling, data minimization, retention, consent

  • Supply Chain Security - SBOM, dependency pinning, artifact signing

  • Safety Evaluation - Red teaming, eval sets, incident readiness

Advanced Patterns:

  • LLM & RAG Production Patterns - Prompt management, safety, cost optimization, caching, monitoring

  • Multi-Region Patterns - Multi-region deployment, data residency, disaster recovery, rollback

  • Online Evaluation Patterns - A/B testing, shadow deployments, feedback loops, automated retraining

  • AgentOps Patterns - AI agent observability, session replay, cost tracking, multi-agent debugging

  • Edge MLOps Patterns - TinyML, federated learning, OTA updates, device-aware CI/CD

  • Cost Management & FinOps - ML/LLM cost modeling, budget guardrails, chargeback, cloud optimization

  • Experiment Tracking Patterns - MLflow/W&B patterns, experiment organization, artifact management, team workflows

  • Automated Retraining Patterns - Trigger strategies, validation gates, safe rollout, canary retraining pipelines

Templates

Use these as copy-paste starting points for production artifacts:

Data Ingestion (dlt)

For loading data into warehouses and pipelines:

  • dlt basic pipeline setup - Install, configure, run basic extraction and loading

  • dlt REST API sources - Extract from REST APIs with pagination, authentication, rate limiting

  • dlt database sources - Replicate from PostgreSQL, MySQL, MongoDB, SQL Server

  • dlt incremental loading - Timestamp-based, ID-based, merge/upsert patterns, lookback windows

  • dlt warehouse loading - Load to Snowflake, BigQuery, Redshift, Postgres, DuckDB

Use dlt when:

  • Loading data from APIs (Stripe, HubSpot, Shopify, custom APIs)

  • Replicating databases to warehouses

  • Building ELT pipelines with incremental loading

  • Managing data ingestion with Python

For SQL transformation (after ingestion), use:

→ ai-ml-data-science skill (SQLMesh templates for staging/intermediate/marts layers)

Deployment & Packaging

  • Deployment & MLOps template - Complete MLOps lifecycle, model registry, promotion workflows

  • Deployment readiness checklist - Go/No-Go gate, monitoring, and rollback plan

  • API service template - Real-time REST/gRPC API with FastAPI, input validation, rate limiting

  • Batch scoring pipeline template - Orchestrated batch inference with Airflow/Dagster, validation, backfill

Monitoring & Operations

  • Monitoring & alerting template - Data/performance/technical metrics, dashboards, SLO definition

  • Drift detection & retraining template - Automated drift detection, retraining triggers, promotion pipelines

  • Incident runbook template - Failure mode playbooks, diagnosis steps, resolution procedures

Navigation

Resources

  • references/drift-detection-guide.md

  • references/model-registry-patterns.md

  • references/online-evaluation-patterns.md

  • references/monitoring-best-practices.md

  • references/llm-rag-production-patterns.md

  • references/api-design-patterns.md

  • references/incident-response-playbooks.md

  • references/deployment-patterns.md

  • references/data-ingestion-patterns.md

  • references/deployment-lifecycle.md

  • references/feature-store-patterns.md

  • references/multi-region-patterns.md

  • references/agentops-patterns.md

  • references/edge-mlops-patterns.md

  • references/cost-management-finops.md

  • references/experiment-tracking-patterns.md

  • references/automated-retraining-patterns.md

Templates

  • template-dlt-pipeline.md

  • template-dlt-rest-api.md

  • template-dlt-database-source.md

  • template-dlt-incremental.md

  • template-dlt-warehouse-loading.md

  • assets/deployment/template-deployment-mlops.md

  • assets/deployment/deployment-readiness-checklist.md

  • assets/deployment/template-api-service.md

  • assets/deployment/template-batch-pipeline.md

  • assets/ops/template-incident-runbook.md

  • assets/monitoring/template-drift-retraining.md

  • assets/monitoring/template-monitoring-plan.md

Data

  • data/sources.json - Curated external references

External Resources

See data/sources.json for curated references on:

  • Serving frameworks (FastAPI, Flask, gRPC, TorchServe, KServe, Ray Serve)

  • Orchestration (Airflow, Dagster, Prefect)

  • Model registries and MLOps (MLflow, W&B, Vertex AI, Sagemaker)

  • Monitoring and observability (Prometheus, Grafana, OpenTelemetry, Evidently)

  • Feature stores (Feast, Tecton, Vertex, Databricks)

  • Streaming & messaging (Kafka, Pulsar, Kinesis)

  • LLMOps & RAG infra (vector DBs, LLM gateways, safety tools)

Data Lake & Lakehouse

For comprehensive data lake/lakehouse patterns (beyond dlt ingestion), see data-lake-platform:

  • Table formats: Apache Iceberg, Delta Lake, Apache Hudi

  • Query engines: ClickHouse, DuckDB, Apache Doris, StarRocks

  • Alternative ingestion: Airbyte (GUI-based connectors)

  • Transformation: dbt (alternative to SQLMesh)

  • Streaming: Apache Kafka patterns

  • Orchestration: Dagster, Airflow

This skill focuses on ML-specific deployment, monitoring, and security. Use data-lake-platform for general-purpose data infrastructure.

Recency Protocol (Tooling Recommendations)

When users ask recommendation questions about MLOps tooling, verify recency before answering.

Trigger Conditions

  • "What's the best MLOps platform for [use case]?"

  • "What should I use for [deployment/monitoring/drift detection]?"

  • "What's the latest in MLOps?"

  • "Current best practices for [model registry/feature store/observability]?"

  • "Is [MLflow/Kubeflow/Vertex AI] still relevant in 2026?"

  • "[MLOps tool A] vs [MLOps tool B]?"

  • "Best way to deploy [LLM/ML model] to production?"

  • "What feature store should I use?"

Minimal Recency Check

  • Start from data/sources.json and prefer sources with add_as_web_search: true .

  • If web search or browsing is available, confirm at least: (a) the tool’s latest release/docs date, (b) active maintenance signals, (c) a recent comparison/alternatives post.

  • If live search is not available, state that you are relying on static knowledge + data/sources.json , and recommend validation steps (POC + evals + rollout plan).

What to Report

After searching, provide:

  • Current landscape: What MLOps tools/platforms are popular NOW

  • Emerging trends: New approaches gaining traction (LLMOps, GenAI ops)

  • Deprecated/declining: Tools or approaches losing relevance

  • Recommendation: Based on fresh data, not just static knowledge

Related Skills

For adjacent topics, reference these skills:

  • ai-ml-data-science - EDA, feature engineering, modelling, evaluation, SQLMesh transformations

  • ai-llm - Prompting, fine-tuning, evaluation for LLMs

  • ai-agents - Agentic workflows, multi-agent systems, LLMOps

  • ai-rag - RAG pipeline design, chunking, retrieval, evaluation

  • ai-llm-inference - Model serving optimization, quantization, batching

  • ai-prompt-engineering - Prompt design patterns and best practices

  • data-lake-platform - Data lake/lakehouse infrastructure (ClickHouse, Iceberg, Kafka)

Use this skill to turn trained models into reliable services, not to derive the model itself.

Fact-Checking

  • Use web search/web fetch to verify current external facts, versions, pricing, deadlines, regulations, or platform behavior before final answers.

  • Prefer primary sources; report source links and dates for volatile information.

  • If web access is unavailable, state the limitation and mark guidance as unverified.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Security

software-security-appsec

No summary provided by upstream source.

Repository SourceNeeds Review
Security

startup-distribution-audit

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

product-management

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

marketing-visual-design

No summary provided by upstream source.

Repository SourceNeeds Review