Running ASR inference pipeline on multiple GPU's

Kaggle notebook have access to 2 GPU’s.

I created two pipelines, set device = 0, device =1. This loaded the inference model in 2 GPU’s.

However, the inference pipeline ran on 1 GPU, while other GPU is idle.

Looking for pointers to run inference on 2 GPU’s in parallel