Looking inside classification after BERT

This is probably a silly question.
Say I have text that I run through BERT and then use a logistic regression on it to predict positive/negative. How can I see how the prediction was made on a large set of data? What if I use a neural network instead of a logistic regression? I’m curious which words/sequences of words affect the prediction the most.