Multi-GPU inference with LLM produces gibberish

Can any one teach me how to use 2 GPUs to run inference? Accelarator can’t detect my GPUs.