(Distributed Training) KeyError: eval_f1 in QuestionAnsweringTrainer taken from trainer_qa.py in examples
|
|
1
|
1196
|
June 22, 2023
|
AlbertForMaskedLM error- "view size is not compatible..."
|
|
1
|
1673
|
June 22, 2023
|
How to use Whisper from huggingface for ASR
|
|
0
|
541
|
June 21, 2023
|
Integration transformers and hidet for wav2vec2
|
|
0
|
128
|
June 21, 2023
|
Unsup fine tuning embeddings
|
|
0
|
412
|
June 21, 2023
|
Seq2seq evaluation speed is slow
|
|
7
|
3840
|
June 20, 2023
|
Getting pretrained embeddings
|
|
0
|
599
|
June 20, 2023
|
How to get embeddings from Finetuned BERT classification model?
|
|
0
|
408
|
June 20, 2023
|
Trying to get predicted text in fine tuned whisper-base.en (Qunatized onnx)
|
|
0
|
237
|
June 20, 2023
|
Finetuning T5 on SQUADv2 with Seq2SeqTrainer fails
|
|
1
|
424
|
June 20, 2023
|
T5 finetuning metrics not improving
|
|
1
|
345
|
June 20, 2023
|
How to save peft weight every epoch?
|
|
1
|
1453
|
June 20, 2023
|
Skipped batches do not consider distributed training
|
|
0
|
690
|
June 19, 2023
|
Why do the F1 and accuracy scores vary when I run the run_glue.py script from Hugging Face's Transformers library for the BERT-base model on the MNLI task, while using different numbers of GPUs?
|
|
0
|
148
|
June 19, 2023
|
Agents Notebook: While is not supported?
|
|
5
|
587
|
June 19, 2023
|
Changing pooling method in pre-trained models
|
|
0
|
1571
|
June 19, 2023
|
GPU error on LoRA for token classification
|
|
2
|
680
|
June 19, 2023
|
Which method is use HF Trainer with multiple GPU?
|
|
4
|
1564
|
June 19, 2023
|
Trainer) training one batch with multiple GPUs
|
|
0
|
396
|
June 19, 2023
|
Issue with push_to_hub
|
|
0
|
502
|
June 18, 2023
|
Why can't temperature be 0 for GPT2 and GPT-Neo?
|
|
2
|
2678
|
June 18, 2023
|
Report `do_predict` metrics in wandb
|
|
0
|
277
|
June 17, 2023
|
ImportError: Using the Trainer with PyTorch: Seq2SeqTrainingArguments
|
|
4
|
5268
|
June 17, 2023
|
Handle long generation in text generation pipeline
|
|
0
|
510
|
June 16, 2023
|
Error: Fine-tune GPT2 model for question answer task
|
|
1
|
797
|
June 16, 2023
|
XLMProphetNet returning different results when using padding
|
|
1
|
242
|
June 16, 2023
|
How to actually use padding in Lllama Tokenizers
|
|
2
|
4949
|
June 16, 2023
|
Indian stock market
|
|
0
|
210
|
June 16, 2023
|
What is Transformers doing? Why it's so slow?
|
|
0
|
1004
|
June 16, 2023
|
Why run_clm keep asking to download pytest
|
|
3
|
319
|
June 15, 2023
|