Great explanations about the process, so thank you for this effort @patrickvonplaten I started the same process for trying different (also too basic) preprocessing steps. Then, I will try different hyperparams settings if Colab let me do it Even I have Colab Pro, it generally limits my GPU connection and prevents to access it. So, if you have any suggestion about tuning hyperparameters faster or easier, I will be glad to hear.
FYI: In Turkish you can omit the ' character from your vocabulary.(in reference to:this)
It’s usually just used as possessive pronoun(?) similar to Ceyda’s == Ceyda’nın OR separating other type of postfixes from special named entities. Doesn’t effect pronunciation
I just finished reading the blogpost+code… The end results were not that I was expecting much with so little data. Now I’m even more motivated to improve it ~ I have plans
Edit: oops it looks like I clicked the wrong reply button @ozcangundes I know you know Turkish and didn’t need the extra explanation
Hi all! I have just completed my initial experiment with 40 epoch and 2e-3 learning rate (the other hparams are kept the same with @patrickvonplaten demo notebook). I run the code from Kaggle since Colab Pro restricted my GPU usage due to long usage time It tooks more than 9 hours and Kaggle also stopped working However, validation loss and WER results did not change dramatically even though training loss decreases sharply, as seen below.
Also, keep in mind that fine tuning CNN layers does not lead to better results as suggested by the authors and demo file. I gave it a shot but the score (50% WER in test set) was not as good as demo file, too.
Hi ! Thanks for the great tutorial. I fined-tuned a pre-trained model (facebook/wav2vec2-large-xlsr-53-french) and got encouraging results. However there are 2 things I wonder (and I think could maybe help some other fellow readers)
In both your notebook @patrickvonplaten and @ozcangundes screenshots, and also from what I experienced, Validation Loss decreases during the first iterations but then starts going up while Wer keeps going down. Do you know why this is happening and can we be sure that WER is the best metric to track (for early stopping at least) especially as I use a custom LM for prediction once my model is done training?
Which are the main hyperparameters to change to try to improve performance (I have limit GPU time and any insights or educated guess on where to start could help ;)) ?