Issue Running OpenAI Inference on Phi-3

Hi, we’re trying to run use the OpenAI Python library and HF Inference Endpoints to run inference on this model: Saturn-App/generator-phi-test

The model is a fine-tuned version of Phi-3 but it’s not working at the moment. Wondering if it’s a compatibility issue or an issue with our implementation.

This is the error message we’re receiving when running our Python code:

NotFoundError                             Traceback (most recent call last)
<ipython-input-8-4d0884690c14> in <cell line: 11>()
      9 )
     10 
---> 11 chat_completion = client.chat.completions.create(
     12         model="tgi",
     13     messages=[

4 frames
/usr/local/lib/python3.10/dist-packages/openai/_base_client.py in _request(self, cast_to, options, remaining_retries, stream, stream_cls)
   1044 
   1045             log.debug("Re-raising status error")
-> 1046             raise self._make_status_error_from_response(err.response) from None
   1047 
   1048         return self._process_response(

NotFoundError: Not Found

We were hoping to get some guidance on this. Thank you!