Now how can I create a config.json file for this?
Normally, if you save your model using the
.save_pretrained() method, it will save both the model weights and a
config.json file in the specified directory.
Yes, but this is a custom model that I have saved in pytorch style, since it consists of additional layers, is there anyway to generate confg.json file?
You need to subclass it to have the save_pretrained methods available. So instead of
from transformers.modeling_utils import PreTrainedModel class Mean_Pooling_Model(PreTrainedModel):
It will add extra functionality on top of
Thank you I will try this!
Is it possible to generate the configuration file for already trained model , i.e weights stored in normal pytorch model.bin
Use model.config.to_json() method to generate config.json
Did you end up finding a solution to getting a config.json from an already trained model? I’m currently struggling with the same problem
Nope, I was not able to find a proper solution, I ended up writing the config.json manually
You should be able to just call
That only works for models that are transformer native and not nn.Module/pytorch native, sadly.
What is your use-case that you are using Transformers but not Transformers models? If you want to use the HF Trainer alongside with your own PyTorch model, I recommended to subclass the relevant classes, similar to PretrainedModel
And to use your own PretrainedConfig alongside of it.
I have a similar issue where I have my model’s (nn.module) weights and I want to convert it to be huggingface compatible model so that I can use hugging face models (as
.generate). From the discussions I can see that I either have to retrain again while changing (
PreTrained) or to define my config.json file. If I wrote my config.json file what should I do next to load my torch model as huggingface one?
I am not sure from the discussion above, what the solution is. Can someone post their working example please?
I am not sure whether this functionality exists at this moment.
I am trying out the code at GitHub - aws-samples/aws-inferentia-huggingface-workshop: CMP314 Optimizing NLP models with Amazon EC2 Inf1 instances in Amazon Sagemaker .
For inferentia instance, I get a config.json not found in the cloudwatch logs and inference fails. The config.json file is present in the traced model tar.gz file for inferentia.
Please help me resolve this.
P.S the log message — W-9002-model_1-stdout MODEL_LOG - OSError: file /home/model-server/tmp/models/cb9491669c1f44c1a0763e8a62d9368e/config.json not found