-
Notifications
You must be signed in to change notification settings - Fork 0
Open
Labels
Description
This project aims to create a draft of the MLOps considerations for a company that has ~20 to 50 ML models running inference in real time for upwards of 100.000 users monthly.
Assumptions
- All infrastructure is on AWS.
- Assume you have an unlimited budget and can use any AWS service.
- Assume each model runs on a standalone Kubernetes pod or EC2 instance for inference. Each of them is accessed through a REST API built on FastAPI.
- All the Data Engineering has been done to get clean data for any given model to a Redshift DataWarehouse.