Error code 400 when running llama2 on sagemaker endpoint
|
|
1
|
1230
|
July 24, 2023
|
Getting ModelError when trying to interact with deployed fine-tuned (LoRA/PEFT) model via Amazon API Gateway and AWS Lambda
|
|
3
|
1691
|
July 21, 2023
|
Sagemaker serverless endpoint deployment error (Image size greater than support size))
|
|
3
|
1236
|
July 21, 2023
|
Recommend an instance for MPT-7B and MPT-30B inference
|
|
2
|
406
|
July 19, 2023
|
Sagemaker instances do not restart after TGI container crashses
|
|
0
|
379
|
July 17, 2023
|
Emotion Model: Additional inference parameter not processed in Sagemaker inferentia instance
|
|
1
|
278
|
July 17, 2023
|
Deploying a conversational pipeline on AWS
|
|
9
|
4302
|
July 13, 2023
|
Problems in deployment when I configure my own labels
|
|
6
|
2668
|
July 12, 2023
|
Training model file too large and fail to deploy
|
|
3
|
1380
|
July 3, 2023
|
Hitting Deployed Endpoint *Outside* of Notebook
|
|
4
|
1939
|
July 4, 2023
|
FP16 doesn't reduce Trainer Training time
|
|
10
|
1836
|
June 29, 2023
|
Sagemaker parameters via AWS client
|
|
2
|
685
|
June 27, 2023
|
Using of context_window in parameter
|
|
0
|
799
|
June 23, 2023
|
Using S3 as model cache for Huggingface LLM inference DLC on Sagemaker
|
|
1
|
3974
|
June 21, 2023
|
HuggingFace Inference containers are not available in GovCloud (US-East)
|
|
2
|
590
|
June 20, 2023
|
OutOfMemoryError: CUDA out of memory while trying to replicate this notebook on sagemaker: https://github.com/huggingface/notebooks/blob/main/sagemaker/24_train_bloom_peft_lora/sagemaker-notebook.ipynb
|
|
4
|
1687
|
June 16, 2023
|
How return custom inference in AWS SageMaker without clone the repo?
|
|
0
|
629
|
June 13, 2023
|
How to pass pipeline parameters when using sagemaker DataSerializer?
|
|
14
|
4174
|
June 9, 2023
|
Deploying Open AI’s whisper on Sagemaker for audio streaming
|
|
2
|
1524
|
June 9, 2023
|
Payload too large for Async Inference on Sagemaker
|
|
8
|
2409
|
June 9, 2023
|
Loading inference.py separately from model.tar.gz
|
|
4
|
1868
|
June 5, 2023
|
Can a diffuser pipeline run on multiple GPUs?
|
|
2
|
1238
|
May 31, 2023
|
Unclear documentation using CLIP on Sagemaker for inference
|
|
1
|
1239
|
May 30, 2023
|
Databricks models deployment to sagemaker are not working
|
|
6
|
1116
|
May 24, 2023
|
Impossible to use flan-t5-xxl in a batch-transform job
|
|
3
|
1150
|
May 23, 2023
|
InternalServerException from bart model created from s3
|
|
1
|
391
|
May 22, 2023
|
CUDA error for inference on GPU instance
|
|
2
|
766
|
May 16, 2023
|
BERT inference with Hugging Face Transformers and AWS Inferentia
|
|
0
|
530
|
May 10, 2023
|
HuggingFace PatentSBERTa Model support in EMR and PySpark
|
|
0
|
400
|
May 8, 2023
|
Errors: Batch transform on fine-tuned models
|
|
4
|
1577
|
May 4, 2023
|