I downloaded model to my local PC and saved it using the following codes.
Codes:
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "beomi/Llama-3-Open-Ko-8B"
model = AutoModelForCausalLM.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
model.save_pretrained("./path/to/model")
tokenizer.save_pretrained("./path/to/model")
Results:
Downloading shards: 100%|██████████| 6/6 [02:50<00:00, 28.42s/it]
Loading checkpoint shards: 100%|██████████| 6/6 [00:06<00:00, 1.01s/it]
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
The actual model files uploaded in huggingface(beomi/Llama-3-Open-Ko-8B at main) has 6 model shards and each of model.safetensors filesize does not exceed 3GB,
while the files in my local PC does not match in the filesize or neither in the number of files.
Can anyone explain this situation or the way to solve the problem?
Thank you in advance