How do I increase max_new_tokens

@gkrishnan I’m late to the post but you can always manually pass in the model/pipeline:

from transformers import pipeline, AutoModelForCausalLM, AutoTokenizer
from langchain.llms import HuggingFacePipeline

model = AutoModelForCausalLM.from_pretrained(model_path)
tokenizer = AutoTokenizer.from_pretrained(model_path)
gen = pipeline('text-generation', model=model, tokenizer=tokenizer, max_new_tokens=200)

llama_llm = HuggingFacePipeline(pipeline=gen)