Kaggle notebook have access to 2 GPU’s.
I created two pipelines, set device = 0, device =1. This loaded the inference model in 2 GPU’s.
However, the inference pipeline ran on 1 GPU, while other GPU is idle.
Looking for pointers to run inference on 2 GPU’s in parallel