How can I load an LLM in 4-bits

I am working with this PEFT example of finetuning a model with 4 bits.

Although I use load_in_4bit=True, I get a log message as follows.

Detected 8-bit loading: activating 8-bit loading for this model

Can someone please explain this to me?

My GPU is Tesla T4 16Gb