I want to monitor my model predictions on validation set not only using metrics but also on few examples by inspecting quality of individual examples.
I am struggling with two aspects.
-
How to get only fixed subset of examples to be logged?
Can I use more than one evaluation dataset and dedicate one to be logged to wandb? -
Where the logging code should live?
In Trainer.Callback? In a fake metric?
Thank you
Ondra