<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>DevOps on StackSimplify | DevOps &amp; Cloud Education by Kalyan Reddy</title><link>https://stacksimplify.com/tags/devops/</link><description>Recent content in DevOps on StackSimplify | DevOps &amp; Cloud Education by Kalyan Reddy</description><generator>Hugo -- gohugo.io</generator><language>en-us</language><lastBuildDate>Tue, 14 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://stacksimplify.com/tags/devops/index.xml" rel="self" type="application/rss+xml"/><item><title>5 Questions to Ask Before Every ML Model Deployment</title><link>https://stacksimplify.com/blog/ml-deployment-checklist/</link><pubDate>Tue, 14 Apr 2026 00:00:00 +0000</pubDate><guid>https://stacksimplify.com/blog/ml-deployment-checklist/</guid><description>A data scientist hands you a model.pkl and says &amp;ldquo;deploy this.&amp;rdquo;
What do you ask?
Most engineers jump straight to containers and endpoints. But the questions that save you at 2 AM are the ones you ask before deployment, not during an incident.
The Checklist # Question Why It Matters 1 What input will break it? Models return garbage confidently on bad input 2 What&amp;rsquo;s the rollback plan? &amp;ldquo;Redeploy the old one&amp;rdquo; is not a plan 3 How do we know it&amp;rsquo;s broken?</description></item><item><title>DevOps Thinking Applied to MLOps: 5 Essential Tools</title><link>https://stacksimplify.com/blog/devops-thinking-mlops-tools/</link><pubDate>Tue, 14 Apr 2026 00:00:00 +0000</pubDate><guid>https://stacksimplify.com/blog/devops-thinking-mlops-tools/</guid><description>If you&amp;rsquo;re a DevOps engineer and a data scientist has ever handed you a model.pkl and said &amp;ldquo;deploy this&amp;rdquo;, you know the feeling.
Where did this come from? What data trained it? Which version is this? How do I scale it?
Here&amp;rsquo;s what I&amp;rsquo;ve learned after months building MLOps pipelines: these aren&amp;rsquo;t new problems. We&amp;rsquo;ve already solved them in DevOps. The tools are different, but the thinking is identical.
The Mental Model: Same Problems, Different Artifacts Every MLOps challenge maps directly to a DevOps pattern you already understand:</description></item><item><title>DVC: Git for Your ML Training Data</title><link>https://stacksimplify.com/blog/dvc-data-version-control/</link><pubDate>Tue, 14 Apr 2026 00:00:00 +0000</pubDate><guid>https://stacksimplify.com/blog/dvc-data-version-control/</guid><description>You version code with Git. What about your model training data?
If you&amp;rsquo;ve ever asked &amp;ldquo;Which dataset trained this model?&amp;rdquo; or &amp;ldquo;Can we reproduce last month&amp;rsquo;s model exactly?&amp;rdquo;, you need DVC.
What DVC Solves Problem Without DVC With DVC Which dataset trained this model? &amp;ldquo;Check the shared drive, maybe?&amp;rdquo; git log shows exact data version Someone changed the training data No history, no diff dvc diff shows exactly what changed Reproduce last month&amp;rsquo;s model Impossible git checkout + dvc checkout Your Weekend Starter Six commands.</description></item><item><title>Feature Stores: The Package Registry for ML Features</title><link>https://stacksimplify.com/blog/feature-stores-ml/</link><pubDate>Tue, 14 Apr 2026 00:00:00 +0000</pubDate><guid>https://stacksimplify.com/blog/feature-stores-ml/</guid><description>Your training pipeline computes &amp;ldquo;average transaction amount&amp;rdquo; as the mean of the last 30 days. Your inference API computes it as the mean of the last 7 days.
Same feature name. Different values. Your model is silently wrong.
This is training-serving skew. The number one silent killer of ML models in production.
The Problem ML features get computed in two places:
Context How Features Are Computed Problem Training Batch job on historical data, saved to CSV Code written by data scientist Serving API computes on the fly per request Different code, different logic Two separate implementations.</description></item><item><title>ML Retraining Pipelines: From Drift Alert to Production Model</title><link>https://stacksimplify.com/blog/ml-retraining-pipelines/</link><pubDate>Tue, 14 Apr 2026 00:00:00 +0000</pubDate><guid>https://stacksimplify.com/blog/ml-retraining-pipelines/</guid><description>Your drift detector triggered an alert. Now what?
Most teams freeze. The runbook says &amp;ldquo;retrain the model.&amp;rdquo; Nobody knows how. Monitoring without a retraining pipeline is like alerting without a runbook.
The Retraining Spectrum Level Trigger Best For Manual Data scientist retrains in a notebook Small teams, low-risk models Scheduled Cron job retrains every week/month Predictable drift patterns Triggered Drift detector kicks off pipeline automatically High-value models Most teams should start with manual.</description></item><item><title>MLflow in 60 Seconds: The Complete ML Model Lifecycle</title><link>https://stacksimplify.com/blog/mlflow-model-lifecycle/</link><pubDate>Tue, 14 Apr 2026 00:00:00 +0000</pubDate><guid>https://stacksimplify.com/blog/mlflow-model-lifecycle/</guid><description>How does an ML model actually get from training to production?
If you&amp;rsquo;re a DevOps engineer stepping into MLOps, MLflow is the first tool you need to understand. It handles the entire lifecycle: tracking experiments, versioning models, and serving them in production.
The 5-Step Lifecycle Here&amp;rsquo;s the full journey of a model, from code to production.
Step What Happens DevOps Analogy Experiment Write training code, MLflow creates a &amp;ldquo;run&amp;rdquo; Starting a CI build Run Logs parameters, metrics, model files Build artifacts + test results Model Best run registered to Model Registry Pushing image to Container Registry Registry Versions (v1, v2, v3) with aliases (@champion, @candidate) Image tags (:latest, :staging, :prod) Serving API loads models:/fraud-detector@champion K8s Deployment pulling :prod tag Step 1: Experiment You write training code and run it.</description></item><item><title>MLOps for DevOps Engineers</title><link>https://stacksimplify.com/blog/mlops-series/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://stacksimplify.com/blog/mlops-series/</guid><description/></item></channel></rss>