Skip to content

ML in Production

Best practices for building real world machine learning systems

  • Newsletter
  • Courses
    • Build, Deploy, and Monitor ML with Amazon SageMaker
  • About

Deploying Models on AWS SageMaker – Part 1 Architecture

A few weeks ago I had the chance to speak to the AI and Data Science Fellows at the Insight Data Science program here in New York City. The fellows in the audience were finishing up their machine learning final… Read More
Author LuigiPosted on July 8, 2019July 19, 2020Categories Deployment1 Comment on Deploying Models on AWS SageMaker – Part 1 Architecture

Kubernetes Services for Machine Learning

Kubernetes Services for Machine Learning
In my previous Kubernetes for Machine Learning post, we used a Kubernetes Deployment to build a REST API to serve a trained machine learning model. In that setup, issuing requests to generate predictions was only possible from within our Kubernetes… Read More
Author LuigiPosted on June 24, 2019June 23, 2019Categories Deployment, Inference, KubernetesLeave a comment on Kubernetes Services for Machine Learning

The Ultimate Guide to Model Retraining

The Ultimate Guide to Model Retraining
Machine learning models are trained by learning a mapping between a set of input features and an output target. Typically, this mapping is learned by optimizing some cost function to minimize prediction error. Once the optimal model is found, it’s… Read More
Author LuigiPosted on June 10, 2019March 21, 2020Categories DeploymentTags Monitoring, Retraining1 Comment on The Ultimate Guide to Model Retraining

Kubernetes Deployments for Machine Learning

Kubernetes Deployments for Machine Learning
Suppose your data science team has deployed a couple of batch machine learning processes on Kubernetes. You’ve successfully used Kubernetes Jobs to deploy model training and you’ve scheduled daily batch inference tasks using CronJobs. But now you’re tasked with serving… Read More
Author LuigiPosted on June 3, 2019June 2, 2019Categories Deployment, Inference, KubernetesLeave a comment on Kubernetes Deployments for Machine Learning

Kubernetes CronJobs for Machine Learning

Kubernetes CronJobs for Machine Learning
In my previous post we discussed how to leverage Kubernetes Jobs to perform common production machine learning tasks such as model training and batch inference. Jobs allow us to reliably run batch processes in a fault tolerant way. Even if… Read More
Author LuigiPosted on May 27, 2019May 25, 2019Categories Deployment, KubernetesLeave a comment on Kubernetes CronJobs for Machine Learning

Kubernetes Jobs for Machine Learning

Kubernetes Jobs for Machine Learning
In my previous post I introduced Kubernetes Pods, the basic building block of the Kubernetes ecosystem. In that post I discussed what a Pod is, how it fits into the Kubernetes system, and how to create, view, and delete a… Read More
Author LuigiPosted on May 20, 2019May 19, 2019Categories Deployment, Inference, Kubernetes, TrainingLeave a comment on Kubernetes Jobs for Machine Learning

How to Use Kubernetes Pods for Machine Learning

How to Use Kubernetes Pods for Machine Learning
One of the best parts of being a Data Scientist is the dynamic nature of the job. You’ll likely spend a majority of your time feature engineering, building models, or running experiments. But depending on your role, you may also… Read More
Author LuigiPosted on May 6, 2019May 5, 2019Categories KubernetesLeave a comment on How to Use Kubernetes Pods for Machine Learning

An Introduction to Kubernetes for Data Scientists

An Introduction to Kubernetes for Data Scientists
In the final piece of my Docker for Machine Learning blog series, I implemented a system for exposing machine learning models as a REST API running in a Docker container. With the container running, I was able to generate predictions… Read More
Author LuigiPosted on April 29, 2019April 29, 2019Categories Kubernetes2 Comments on An Introduction to Kubernetes for Data Scientists

Tracking Machine Learning Metadata with Sacred Library

Tracking Machine Learning Metadata with Sacred Library
A few weeks ago I wrote about why storing metadata is critical for the machine learning process. Since building machine learning models is an iterative process, often involving multiple people and a diverse set of tools, we need the ability… Read More
Author LuigiPosted on April 22, 2019April 21, 2019Categories Experiments, MetadataLeave a comment on Tracking Machine Learning Metadata with Sacred Library

How Data Leakage Impacts Machine Learning Models

How Data Leakage Impacts Machine Learning Models
The silver bullet. A feature that led to AUC increasing from .6 to .8. After working on feature engineering for several months, I thought I had finally cracked the code and created a feature that pushed my machine learning model… Read More
Author LuigiPosted on April 16, 2019April 16, 2019Categories Training2 Comments on How Data Leakage Impacts Machine Learning Models

Posts pagination

Previous page Page 1 … Page 4 Page 5 Page 6 Next page

Mission

My goal is to help data scientists, ML engineers, and AI product managers, build and operate machine learning systems in production.

Learn more about why I started MLinProduction.

Enroll in my online course

Sagemaker Course Banner

Recent Posts

  • Driving Experimentation Forward through a Working Group (Experimentation Program Series: Guide 03)
  • What is an Experimentation program and Who is Involved? (Experimentation Program Series: Guide 02)
  • Building An Effective Experimentation Program (Experimentation Program Series: Guide 01)
  • Lessons Learned from Writing Online
  • Newsletter #087
  • Newsletter
  • Courses
    • Build, Deploy, and Monitor ML with Amazon SageMaker
  • About
ML in Production Disclaimer