In this video, I show you how to deploy Transformer models straight from the Hugging Face hub to managed infrastructure on AWS, in just a few clicks.
Starting from a model that I already trained for image classification, I first deploy an endpoint protected by Hugging Face token authentication. Then, I deploy a second endpoint in a private subnet, and I show you how to access it securely from your AWS account thanks to AWS PrivateLink.
- Model: https://huggingface.co/juliensimon/autotrain-food101-1471154053
- Inference Endpoints: https://huggingface.co/inference-endpoints
- Inference Endpoints documentation: https://huggingface.co/docs/inference-endpoints/index
- AWS PrivateLink documentation: https://docs.aws.amazon.com/vpc/latest/privatelink/concepts.html