How to configure ONNX models from Hugging Face to use model options in C++?
|
|
0
|
486
|
November 10, 2023
|
Is a wheel to be released with the 1.14.0 release?
|
|
1
|
366
|
November 7, 2023
|
Donut fine tuning question
|
|
0
|
1548
|
October 16, 2023
|
Optimum export ONNX failure
|
|
0
|
674
|
September 30, 2023
|
Improving Whisper for Inference
|
|
11
|
3773
|
September 20, 2023
|
Order between optimization and quantization
|
|
1
|
506
|
September 19, 2023
|
Optimum-Cli [ Task Manager Error ]
|
|
1
|
696
|
September 18, 2023
|
Custom data preparation for LayoutLM model
|
|
1
|
1089
|
September 18, 2023
|
Static quantization of gpt2-style models with ORTQuantizer
|
|
3
|
858
|
September 18, 2023
|
How to Prune Transformer based Model?
|
|
2
|
5248
|
August 25, 2023
|
How to ensure that while running with llama2-70B, we use parallelism?
|
|
11
|
1580
|
August 22, 2023
|
How to load checkpoint shards with gaudi instead of cpu?
|
|
1
|
948
|
August 21, 2023
|
Error while Trying to run inference using gaudi on a finetuned llama2 model using habana repo
|
|
9
|
653
|
August 21, 2023
|
ORT CLI vs. Programmatic
|
|
12
|
1265
|
August 17, 2023
|
4 Bit quantization
|
|
4
|
546
|
August 11, 2023
|
Static quantization of activations for transformers
|
|
2
|
1573
|
August 11, 2023
|
Export a BetterTransformer to ONNX
|
|
3
|
2765
|
August 11, 2023
|
Exporting model wav2vec2 not supported?
|
|
3
|
1188
|
August 10, 2023
|
BLIP-2 on Optimum
|
|
4
|
1263
|
July 21, 2023
|
No module named 'optimum.neuron'; 'optimum' is not a package
|
|
2
|
2100
|
July 21, 2023
|
Dmesg: read kernel buffer failed: Operation not permitted :- Running gaudi-enabled habana model inference on kubernetes cluster
|
|
1
|
3253
|
July 13, 2023
|
InvalidArgument: [ONNXRuntimeError] : 2 : INVALID_ARGUMENT : Unexpected input data type. Actual: (tensor(int32)) , expected: (tensor(int64)
|
|
2
|
4700
|
July 9, 2023
|
Static Quantization with Own dataset
|
|
3
|
908
|
July 1, 2023
|
Is there a way to include the text_projection and/or embedding normalization in an optimum-optimized CLIPTextModelWithProjection?
|
|
8
|
1247
|
July 1, 2023
|
Support onnx opset 9 for T5 & GPT_neox
|
|
1
|
412
|
July 1, 2023
|
How to use the cache_dir along with optimum-cli export
|
|
5
|
805
|
June 16, 2023
|
ONNX Flan-T5 Model OOM on GPU
|
|
2
|
2553
|
June 15, 2023
|
Potential Memory Leak for ORTModelForCausalLM with TensorRT Providor
|
|
4
|
893
|
June 2, 2023
|
HOw to make optimum make use of all available GPUs?
|
|
7
|
3404
|
June 1, 2023
|
ONNX vs. Apache TVM
|
|
0
|
781
|
June 1, 2023
|