Using S3 as model cache for Huggingface LLM inference DLC on Sagemaker

We release a blog post on how to do this: Securely deploy LLMs inside VPCs with Hugging Face and Amazon SageMaker

1 Like