Inference on constrained devices

Hi there,
I am looking for any resources or any previous work on getting huggingface models to run inference on constrained devices. Since I read on your DistilGPT2 page that it “Runs smoothly on an iPhone 7.” I’ve been curious.

Has anyone managed to get inference working on something like an RPi?

Many Thanks,