Inference on constrained devices

Hi there,
I am looking for any resources or any previous work on getting huggingface models to run inference on constrained devices. Since I read on your DistilGPT2 page that it “Runs smoothly on an iPhone 7.” I’ve been curious.

Has anyone managed to get inference working on something like an RPi?

Many Thanks,
Chris