Cluster tried to load adapter_config.json even though model is not PEFT-based model
|
|
1
|
374
|
April 5, 2024
|
Datetime parsing error when calling get_inference_endpoint
|
|
0
|
90
|
April 2, 2024
|
Get HF token into custom handler
|
|
0
|
128
|
March 29, 2024
|
HF Inference Endpoints don't finish Initializing
|
|
0
|
132
|
March 28, 2024
|
Is there an response length limit for the inference API?
|
|
0
|
180
|
March 28, 2024
|
[RESOLVED] Recommended way to use guidance on an Inference Endpoint?
|
|
1
|
176
|
March 27, 2024
|
Multiple Requests to HuggingFace InferenceEndpoints are not working with custom Docker deployment. :-(
|
|
0
|
204
|
March 26, 2024
|
Charge when endpoint’s initializing
|
|
0
|
119
|
March 23, 2024
|
Pricing for Huggingface Endpoint
|
|
5
|
1620
|
March 22, 2024
|
Q: How to query Inference Endponts for Feature Extraction task
|
|
1
|
132
|
March 22, 2024
|
Hosting Mistral 7b quantized 4bit
|
|
2
|
447
|
March 19, 2024
|
How do I add a stop token for Inference Endpoints?
|
|
0
|
157
|
March 19, 2024
|
Error: Command 'apt install -y tesseract-ocr' returned non-zero exit status 100
|
|
0
|
134
|
March 19, 2024
|
Inference Endpoint not stable
|
|
3
|
239
|
March 18, 2024
|
Error invoking DialoGPT-large via serverless inference endpoint - can only concatenate str (not "dict") to str"
|
|
3
|
468
|
March 14, 2024
|
ShardCannotStart Error when launching a dedicated endpoint
|
|
0
|
242
|
March 13, 2024
|
Stopping criteria
|
|
2
|
279
|
March 12, 2024
|
API Endpoint not working as expected
|
|
1
|
277
|
March 10, 2024
|
Inference endpoint "failed" and then "deleted"
|
|
1
|
261
|
March 8, 2024
|
Inference Endpoint Fails to Start
|
|
16
|
1832
|
February 9, 2024
|
Inference Endpoint not starting on HTTP request
|
|
2
|
155
|
March 6, 2024
|
Autoscaling on inference endpoints not initializing from 0 replicas
|
|
0
|
140
|
March 6, 2024
|
Convert PyTorch Model to Hugging Face model
|
|
0
|
377
|
March 5, 2024
|
Trouble returning audio from Interference endpoints
|
|
2
|
297
|
February 28, 2024
|
50 ms inference, 500 ms latency
|
|
0
|
137
|
February 27, 2024
|
Cannot log in to inference endpoint webapp
|
|
0
|
243
|
February 23, 2024
|
Cannot run large models using API token
|
|
5
|
4208
|
February 22, 2024
|
Getting the "Test your endpoint" playground code
|
|
0
|
112
|
February 22, 2024
|
How can I create an endpoint for a model but with a different config?
|
|
1
|
154
|
February 21, 2024
|
Full log history endpoint
|
|
0
|
123
|
February 17, 2024
|