Error Using Pydantic with LangChain and local model by Hugging Face for Structured Output
|
|
1
|
1103
|
October 20, 2024
|
How to parallel infer multiple input sentences with beam search = 4?
|
|
0
|
24
|
October 20, 2024
|
Trying to understand system prompts with Llama 2 and transformers interface
|
|
9
|
46173
|
October 19, 2024
|
Unable to achieve better performance with transformer than LSTM
|
|
0
|
419
|
October 17, 2024
|
@huggingface/transformers library won't run under node-alpine
|
|
0
|
44
|
October 17, 2024
|
Summary of the meeting in Indonesian
|
|
0
|
13
|
October 17, 2024
|
Machine learning
|
|
1
|
40
|
October 17, 2024
|
First time to AI - apps. Do I need a GPU in order to run a model using transformers?
|
|
1
|
300
|
October 17, 2024
|
Fine-tuning LLM
|
|
0
|
68
|
October 16, 2024
|
How to run single-node, multi-GPU training with HF Trainer?
|
|
5
|
15300
|
October 16, 2024
|
Trainer leaked memory?
|
|
1
|
785
|
October 15, 2024
|
Modifying ViT to include 4th channel
|
|
2
|
515
|
October 15, 2024
|
Loading pre-trained models with AddedTokens
|
|
2
|
789
|
October 14, 2024
|
Problem with pushing quantized model to hub
|
|
3
|
313
|
October 14, 2024
|
How to use Cache with message API
|
|
0
|
16
|
October 13, 2024
|
How do I do inference using the GPT models on TPUs?
|
|
5
|
2519
|
October 13, 2024
|
Getting token probabilities of a caption given an image from BLIP2
|
|
4
|
479
|
October 13, 2024
|
Why is BCELoss used for multi-label classification?
|
|
4
|
453
|
October 12, 2024
|
Best practice to train LLMs on long sequences?
|
|
0
|
60
|
October 12, 2024
|
Different Trainers, when to use which?
|
|
1
|
1684
|
October 12, 2024
|
Gradients in Data Collator cause Memory Leak
|
|
4
|
186
|
October 12, 2024
|
Question About the Transformer Concept
|
|
0
|
10
|
October 12, 2024
|
If I use llama 70b and 7b for speculative decoding, how should I put them on my multiple gpus in the code
|
|
0
|
48
|
October 11, 2024
|
Transformers cache not loading from a new vm
|
|
6
|
151
|
October 11, 2024
|
Questions about vocab size, decoder start token, padding token, and appropriate config for custom seq2seq transformer model without any tokenizer
|
|
0
|
55
|
October 11, 2024
|
Any Multi Modal LLMs that take direct pdf + text as input?
|
|
2
|
2128
|
October 10, 2024
|
Training CausalLM to imitate Seq2SeqModel
|
|
2
|
678
|
October 10, 2024
|
Seq2seq padding
|
|
1
|
75
|
October 10, 2024
|
Whisper for Audio Classification
|
|
3
|
2976
|
October 9, 2024
|
Should I Include Poet Information as a Feature in LLM Training with 3,356 Unique Poets?
|
|
0
|
38
|
October 9, 2024
|