Here is the exception and code.
Expected object of device type cuda but got device type cpu for argument #3 ‘index’ in call to _th_index_select
from transformers import AutoTokenizer, AutoModelForQuestionAnswering, pipeline import torch BERT_DIR = "savasy/bert-base-turkish-squad" tokenizer = AutoTokenizer.from_pretrained(BERT_DIR) model = AutoModelForQuestionAnswering.from_pretrained(BERT_DIR) nlp=pipeline("question-answering", model=model, tokenizer=tokenizer) def infer(question,corpus): try: ans = nlp(question=question, context=corpus) return ans["answer"], ans["score"] except: ans = None pass return None, 0