Looking for Pre-trained Model for Image Categorization (Screenshots, Photos, Scans, etc.)
|
|
2
|
260
|
April 3, 2024
|
Extract information from nested tables and charts (Bar/Pie charts)
|
|
0
|
75
|
April 3, 2024
|
Your request to access this repo has been successfully submitted, and is pending a review from the repo's authors
|
|
8
|
15429
|
April 1, 2024
|
Using an encoder-decoder model for Recognizing Textual Entailment (GLUE task)
|
|
0
|
64
|
March 31, 2024
|
Help Needed: Fine-Tuned Model for Georgian Language Not Generating Text
|
|
0
|
75
|
March 31, 2024
|
Llama2 pad token for batched inference
|
|
7
|
10396
|
March 31, 2024
|
Llama/Mistral Finetuning for Inference API
|
|
0
|
85
|
March 30, 2024
|
8 bit precision error
|
|
0
|
153
|
March 30, 2024
|
DistillGpt2 only predicts endoftext if context is full
|
|
0
|
50
|
March 30, 2024
|
Using mlx lora.py with llama-2-13b and mixtral-8x7b
|
|
0
|
131
|
March 30, 2024
|
Don't average the loss
|
|
1
|
392
|
March 30, 2024
|
Hermes 2's secret origin story? 🧐
|
|
0
|
84
|
March 29, 2024
|
Fine tuning existing hugging face model on new dataset (text to sql task)
|
|
0
|
995
|
December 6, 2023
|
Incomplete/ partial response generation
|
|
3
|
871
|
March 27, 2024
|
How to add noise to the intermediate layer of huggingface bert model?
|
|
0
|
78
|
March 27, 2024
|
Best model for generating multiple choice questions
|
|
0
|
91
|
March 27, 2024
|
Which Transformers model is suitable for Video to text and summarize the text?
|
|
0
|
82
|
March 25, 2024
|
Finetuning Bert for Question answering task without context
|
|
0
|
132
|
March 25, 2024
|
Find models by size
|
|
3
|
425
|
March 24, 2024
|
How would you train a model for hard/soft skill detection based on a taxonomy?
|
|
3
|
126
|
March 24, 2024
|
MCQA Model Underfitting the Training Data
|
|
0
|
70
|
March 22, 2024
|
NameError: name 'feature_extractor' is not defined
|
|
4
|
183
|
March 23, 2024
|
Multi-page Document Classification
|
|
3
|
1205
|
March 22, 2024
|
Find LLM to run on single gpu with only 8 GB ram
|
|
10
|
1265
|
March 22, 2024
|
codellama/CodeLlama-70b-Instruct-hf TGI server out-of-memory error in H100
|
|
2
|
143
|
March 22, 2024
|
yarn-mistral-7b-128k.Q8_0.gguf response seems out of control
|
|
0
|
88
|
March 21, 2024
|
torch.cuda.OutOfMemoryError for CodeLlama models in H100 single GPU inference
|
|
2
|
159
|
March 21, 2024
|
What is best LLM model for text classification using few shot learning?
|
|
0
|
137
|
March 21, 2024
|
Which PR to use for safetensors?
|
|
1
|
84
|
March 20, 2024
|
Llama2 prompt template for finetuning on text summaraization/generation
|
|
0
|
129
|
March 20, 2024
|