Need help on wav2vec 2.0 models training

Hello guys, I’m using transformers library and I want to build speech recognition systems base on wav2vec 2.0

I have different problems

  1. Base on the exemple here Wav2Vec2 — transformers 4.7.0 documentation
    I have tried to pretrain a model. But I have seen in the github on fairseq wav2vec that the loss object at the end, I can call a backward method on, but here I can.

  2. how can I perform batch training in wav2vec.

Guys please I need help, even another documentation or examples.

https://github.com/huggingface/blog/blob/master/fine-tune-wav2vec2-english.md

try this

1 Like

Thanks you