One of the features of optuna
is its support for asynchronous parallelization of trials across multiple devices (see its doc) . But from my experiences of using trainer.hyperparameter_search()
, it seems that different trials are executed one after another. So even though I do have access to multiple devices, I could not leverage them for parallel HPO.
I am wondering if it is possible to use the parallelization feature of optuna
. Any input is appreciated.