Colab's session crashed after using all available RAM when loading falcon-7B
|
|
2
|
1207
|
October 26, 2023
|
Double expected memory usage
|
|
1
|
1425
|
August 17, 2022
|
GPU quota exceeded even when using access token from PRO
|
|
1
|
1400
|
June 7, 2024
|
We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like bigcode/starcoderbase-1b is not the path to a directory containing a file named config.json
|
|
7
|
12335
|
December 13, 2024
|
Device_map="auto" with error: Expected all tensors to be on the same device
|
|
7
|
6914
|
January 5, 2025
|
I can't concatenate_datasets because features are not sorted. How do I sort it?
|
|
3
|
5497
|
August 11, 2021
|
How to load pretrained model with custom model layers
|
|
2
|
1111
|
September 12, 2023
|
How to log predictions from evaluation set after each Trainer validation to wandb?
|
|
2
|
1110
|
March 14, 2024
|
Huggingface-cli login hangs
|
|
3
|
3028
|
July 8, 2024
|
'Type Error: list object cannot be interpreted as integer' while evaluating a summarization model (seq2seq,BART)
|
|
4
|
8540
|
November 30, 2021
|
LLM model repository file format
|
|
0
|
6012
|
December 4, 2023
|
Tensorflow model.summary() doesn't show detail of TFBertModel
|
|
0
|
1063
|
August 20, 2020
|
ONNX model created with Optimum is not compatible with Transformers.js
|
|
0
|
1058
|
April 5, 2024
|
Using multi GPU with Trainer through Deepspeed, parameters found on cpu
|
|
0
|
1057
|
August 9, 2023
|
HOW TO determine the best threshold for predictions when making inference with a finetune model?
|
|
4
|
8383
|
December 25, 2023
|
BERT Next Sentence Prediction: How to do predictions?
|
|
5
|
7653
|
September 29, 2022
|
Avoiding the usage of HfApiModel and using local model - `smolagents`
|
|
7
|
1177
|
May 2, 2025
|
Error: HTTPSConnectionPool
|
|
0
|
1052
|
June 4, 2023
|
Error: The model weights are not tied. Please use the `tie_weights` method before using the `infer_auto_device` function, even after adding model.tie_weights()
|
|
8
|
11041
|
February 20, 2024
|
AutoModelForCausalLM.from_pretrained gets stuck when loading model from local folder
|
|
2
|
3289
|
January 1, 2025
|
Where in the code does masking of tokens happen when pretraining BERT
|
|
5
|
7307
|
August 17, 2020
|
Pyannote/speaker-diarization - [WinError 1314] A required privilege is not held by the client
|
|
5
|
7269
|
June 22, 2024
|
LLAMA-2 conversation generated responses always empty
|
|
1
|
3981
|
September 21, 2023
|
[Beginner] fine-tune Bart with custom dataset in other language?
|
|
2
|
3248
|
January 22, 2021
|
Download LLM Model
|
|
4
|
14136
|
May 12, 2024
|
Where is the model path on macOS
|
|
1
|
7055
|
August 9, 2022
|
I want to deploy Hugging Face with ONNX in JavaScript for question and answering
|
|
3
|
1577
|
July 27, 2022
|
Trying the inference with model Llama-2-70b-hf on 2 A100 (80g) GPUs but getting errors
|
|
6
|
6679
|
November 28, 2023
|
Apple silicon Installation of Transformers
|
|
4
|
7823
|
April 17, 2021
|
Non shuffle training
|
|
6
|
6573
|
August 26, 2024
|
Encoding sentence pair with BERT cause ValueError: not enough values to unpack (expected 2, got 1)
|
|
1
|
6855
|
November 13, 2022
|
Image data augmentation - ViT
|
|
1
|
1212
|
July 28, 2022
|
How to load a pretrained custom model using `from_pretrained`
|
|
4
|
7637
|
June 21, 2023
|
Purpose of padding and truncating
|
|
7
|
3385
|
August 3, 2020
|
Inference on Multi-GPU/multinode
|
|
4
|
7606
|
January 12, 2023
|
ValueError: Expected input batch_size (4096) to match target batch_size (8)
|
|
3
|
8452
|
April 2, 2023
|
No module named 'accelerate' even when it's installed
|
|
6
|
11272
|
August 4, 2025
|
Proper way to do conditional generation with T5
|
|
1
|
2107
|
January 20, 2023
|
TrainingArguments cosine_with_restarts: how to define the restart count?
|
|
2
|
3011
|
March 21, 2025
|
How to handle very large datasets
|
|
1
|
2029
|
June 12, 2023
|
Modifying Whisper using Domain Specific Attention
|
|
2
|
931
|
June 15, 2025
|
Codebase Embedding
|
|
4
|
2273
|
January 30, 2025
|
I am completely lost on the hugging face site
|
|
6
|
1079
|
April 8, 2025
|
Invalid token or no access to Hugging Face
|
|
3
|
2534
|
May 29, 2025
|
What can cause model.generate (BART) output to be gibberish after fine-tuning?
|
|
3
|
4463
|
August 31, 2020
|
How do I create Datasets from PDF files?
|
|
8
|
1665
|
August 3, 2025
|
New Framework smolagents
|
|
3
|
783
|
January 15, 2025
|
Does setting max_seq_length to a too large number for fine tuning LLM using SFTTrainer affects model training?
|
|
1
|
1950
|
December 6, 2024
|
Trainer.evaluate() with text generation
|
|
5
|
3557
|
December 31, 2021
|
Running LLM on Android
|
|
1
|
3435
|
March 9, 2025
|