Training models for smaller epochs and then continue trianing
|
|
5
|
1304
|
January 16, 2021
|
Few shot text generation with T5 transformers like GPT-3
|
|
3
|
5853
|
January 15, 2021
|
[Announcement] GenerationOutputs: Scores, Attentions and Hidden States now available as outputs to generate
|
|
1
|
4574
|
January 13, 2021
|
Problem while uploading a file
|
|
23
|
4897
|
January 12, 2021
|
Question About Attention Score Computation & Intuition
|
|
1
|
1649
|
January 12, 2021
|
Improvements with SWA
|
|
5
|
3023
|
January 12, 2021
|
Issues running seq2seq distillation
|
|
4
|
862
|
January 11, 2021
|
Seq2Seq-Example does not work on Azure
|
|
2
|
808
|
January 9, 2021
|
Instantiating TransfoXLTokenizer using existing vocab dict
|
|
1
|
281
|
January 8, 2021
|
Parallelize model call for TFBertModel
|
|
3
|
1025
|
January 7, 2021
|
How to train new token embedding to add to a pretrain model?
|
|
1
|
3602
|
January 6, 2021
|
Inverse T5 with output (instead of input) prefix
|
|
2
|
515
|
January 6, 2021
|
How can I download my private model?
|
|
4
|
8494
|
January 5, 2021
|
[Beginner] ClassificationModel Running out of Memory, long training Epochs
|
|
6
|
1798
|
January 4, 2021
|
AttributeError: 'TrainOutput' object has no attribute 'metrics' when finetune custom dataset
|
|
3
|
2501
|
January 4, 2021
|
Easy way to implement annealing temperature softmax
|
|
1
|
1215
|
January 4, 2021
|
Api and parameters change from transofrmers 2.5.1 to 3.5.1 for GPT2
|
|
0
|
238
|
January 4, 2021
|
Model saving results in a small size checkpoint
|
|
1
|
616
|
January 4, 2021
|
Is there a way to return the "decoder_input_ids" from "tokenizer.prepare_seq2seq_batch"?
|
|
5
|
3336
|
December 29, 2020
|
RoBERTa Tokenizer supported characters
|
|
0
|
620
|
December 24, 2020
|
Sampling with FSMTForConditionalGeneration
|
|
1
|
326
|
December 24, 2020
|
Recommend argument values for transformers.generation_utils.GenerationMixin.generate() for summarization and translation tasks?
|
|
1
|
379
|
December 24, 2020
|
Why loss printed by fit() differs from loss using custom loop for huggingface RoBERTa?
|
|
0
|
321
|
December 23, 2020
|
Batch inference using tfserving/kfserving
|
|
0
|
520
|
December 22, 2020
|
About the origin of the model category names in `AutoModelWithLMHead`
|
|
2
|
1531
|
December 21, 2020
|
Unable to import Hugging Face transformers
|
|
2
|
2041
|
December 21, 2020
|
Huggingface Transformer code successfully gets executed on amazon web services but not on other server
|
|
1
|
1362
|
December 20, 2020
|
Transformers Tokenizer on GPU?
|
|
3
|
14727
|
December 17, 2020
|
Streamlit app with PyTorch/HuggingFace Transformers crashes when deployed to Heroku
|
|
0
|
853
|
December 16, 2020
|
Loading Lower Layers of Model
|
|
1
|
2121
|
December 16, 2020
|