Run parallel api inference for QA

How can we handle the parallel request of question-answering using API inference for qa?