TL;DR —
We often get asked whether serverless is the right compute architecture to deploy models. The cost savings touted by serverless seem appealing for ML workloads as for other traditional workloads. However, the special requirements of ML models as related to hardware and resources can cause impediments to using serverless. This blog post talks about how to get started with deploying models on AWS Lambda, along with the pros and cons of using this system for inference. In particular, we use the DistillBERT question and answer model from HuggingFace.
[story continues]
Written by
@conrado
CTO at Verta.ai
Topics and
tags
tags
model-deployment|serverless-framework|machine-learning|mlops|aws-lambda|aws|amazon|devops
This story on HackerNoon has a decentralized backup on Sia.
Transaction ID: 1KAEftnTe5IAZMJhXHql_jPDaJp0jiqY_GlXGpIvBIQ
