I just loaded the meta-llama/llama2-13b-hf model and did a counted the parameters using the numel() fn. It looks like the 13b model has ~6.7b params. I’m see a similar discrepancy with the meta-llama/llama2-7b-hf model as well.
However the model sizes were fine a week ago.
Did the models change?