Generate function returns random words for BartForConditionalGeneration

Hello, I am trying to use BartForConditionalGeneration, but the generate function returns random words. The model I am using is facebook/bart-base and it has not been fine-tuned at all. Here is the code:

device = torch.device('cuda') if torch.cuda.is_available() else torch.device('cpu')
model_name = "facebook/bart-base"
tokenizer = AutoTokenizer.from_pretrained(

model_cfg= BartConfig.from_pretrained(model_name)
model_cfg.max_length = max_length
# model_cfg.use_cuda=True
model_cfg.force_bos_token_to_be_generated = True
model = BartForConditionalGeneration(model_cfg).to(device)

Here is the generation code:

from transformers.models.bart.modeling_bart import shift_tokens_right
ARTICLE_TO_SUMMARIZE ="My friends are cool but they eat too many carbs."   

inputs = tokenizer([ARTICLE_TO_SUMMARIZE],          

summary_ids = model.generate(shift_tokens_right(inputs['input_ids'],tokenizer.bos_token_id,tokenizer.eos_token_id), num_beams=4,max_length=40)
print([tokenizer.decode(g, skip_special_tokens=True) for g in summary_ids])

The output of this is:

[' society society society Steve Steve Steve smoking smoking smoking contiguous contiguous contiguous Canucks Canucks Canucks concurrent concurrent concurrentRandRandRandLessLessLess Providence Providence ProvidenceumatumatumatCombatCombatCombat vicious vicious vicious Kre']

I have been trying to solve different issues with generate like this for nearly the past three day and a half, but have no idea…

Any help is incredibly appreciated

Nvm. I realized I still needed to load the pretrained model and not use the config