I have read this tutorial: Fine-Tune Whisper For Multilingual ASR with 🤗 Transformers
and when i go to the step of
#We can apply the data preparation function to all of our training examples using dataset's .map method common_voice = common_voice.map(prepare_dataset,remove_columns=common_voice.column_names["train"],num_proc=4)
It really take a loonnnnnng time and i must do it every run after shutdown my pc …etc is there any way to make easier not to train them once ! I don’t know how i am new to Transformers
- how can i fine tune the AJX version ?
- I tried to fine tune the model in kaggle but get string error : raise RuntimeError(f"Failed to load audio from {fileobj}")
RuntimeError: Failed to load audio from <_io.BytesIO object at 0x77788b6c0900> - I triedy also to fine tuen it on colab but it get stuck in the step of common_voice.map(…etc) and take a long time without any progress