Hello @philschmid
Issue was not with saving model artifacts on the top level.
Solution: To install additional libraries on the container, libraries which are part of requirements.txt text file needs to be installed using pip in inference script. Within the archive, the HuggingFace container expects all inference code to be inside the code/
directory.
I did not find any article on multi-model endpoint with hugging face and examples in “aws-samples” git repo hence I publish this article on Medium.
Hope that helps everyone. Thanks