Chapter 7 questions

I have a question regarding the fine-tuning of a causalLM model (Llama2) with the trainer. I set the seed 42 but the train and evaluation loss differ across runs when fine-tuning Llama2 on my dataset. This behavior is only observable with Llama2 (I tried it with Mistral and there was the loss always the same). May this problem be related to the model.generation_config file? is there any non-deterministic behavior within the train loop?