Hi, we’re trying to run use the OpenAI Python library and HF Inference Endpoints to run inference on this model: Saturn-App/generator-phi-test
The model is a fine-tuned version of Phi-3 but it’s not working at the moment. Wondering if it’s a compatibility issue or an issue with our implementation.
This is the error message we’re receiving when running our Python code:
NotFoundError Traceback (most recent call last)
<ipython-input-8-4d0884690c14> in <cell line: 11>()
9 )
10
---> 11 chat_completion = client.chat.completions.create(
12 model="tgi",
13 messages=[
4 frames
/usr/local/lib/python3.10/dist-packages/openai/_base_client.py in _request(self, cast_to, options, remaining_retries, stream, stream_cls)
1044
1045 log.debug("Re-raising status error")
-> 1046 raise self._make_status_error_from_response(err.response) from None
1047
1048 return self._process_response(
NotFoundError: Not Found
We were hoping to get some guidance on this. Thank you!