Skip to main content
Presentation

From Training to Serving: Machine Learning Models with Terraform

In this talk you'll see a HashiCorp Terraform-managed architecture built in AWS to handle the full machine learning and natural language processing (NLP) lifecycle.

As machine learning becomes more pervasive across industries, the need to automate the deployment of the required infrastructure becomes even more important. The ability to efficiently and automatically provision infrastructure for modeling training, evaluation, and serving becomes an important component of a successful ML pipeline.

What You'll Learn

In this talk you'll see a HashiCorp Terraform-managed architecture built in AWS to handle the full machine learning lifecycle. Using tools and services such as containers, Amazon EC2, S3, and Lambda, our architecture will facilitate training and evaluating natural language processing (NLP) models.

You'll come away with a working knowledge of how a machine learning pipeline can be constructed and managed with Terraform. Knowledge of NLP is not required and all NLP concepts key to the talk will first be introduced. While the talk will use NLP as an example, the processes described will largely be generic and adaptable to other types of machine learning models. All code presented is available on GitHub.

Speakers: Jeff Zemerick, David Smithbauer

More resources like this one

3/15/2023Presentation

Advanced Terraform techniques

2/3/2023Case Study

Automating Multi-Cloud, Multi-Region Vault for Teams and Landing Zones

2/1/2023Case Study

Should My Team Really Need to Know Terraform?

1/20/2023Case Study

Packaging security in Terraform modules