Global Transformer in Llama 3.2 Vision

There two transformers in the vision encoder. One is called global_transformer and the other transformer.
I see is_gated is different. What is global about the ‘global_transformer’?

self.transformer = MllamaVisionEncoder(config, config.num_hidden_layers, is_gated=False)
self.global_transformer = MllamaVisionEncoder(config, config.num_global_layers, is_gated=True)
Thanks.

1 Like

It seems to be turned on by output_hidden_states and/or output_attentions.