Difference between pinned models and Inference endpoints

Thank you a lot for the answer.

I have a follow-up question:

For Inference API it is said that the model is run on Intel Ice Lake CPU but the instance is not explicitly mentioned. Can you tell me which of the following instances does the Inference API use?