I would like to observe the attention between an input RoI and each word in an input sentence of LXMERT. If a framework that facilitates what I want do exists, please let me know. If not, could you tell me which of the tensors from LXMERT I should watch?
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
How to visualize attention of a large encoder-decoder transformer model that isn't a model on hugging face? | 0 | 2301 | June 28, 2021 | |
How can one visualize the Cross-Attention of a VisionEncoderDecoderModel? | 2 | 1887 | November 7, 2023 | |
How do we use LXMERT for inference? | 2 | 511 | December 25, 2021 | |
Understanding what went wrong in attention | 5 | 1639 | July 31, 2020 | |
Visualizing Attention Maps in SwinV2 | 1 | 3011 | December 29, 2023 |