My understanding from reading the PyTorch XLA documentation is that xmp.spawn()
is used for multi-TPU processing. For single-TPU training, you only need to define device
correctly. The difference is also shown in the PyTorch example code for single-core AlexNet and multi-core AlexNet training.
At this time, I’m interested in just single-TPU execution.