Hi all!
I’m going to release a huge dataset (about 120 GB) of text to the / Datasets which I believe will help the Persian nlp community a lot. I haven’t tried uploading the dataset directly to the hub, it may be possible to do so by git-lfs but I’m not sure about it.
Is there any free hosting service that I can upload the big .txt file into it and then use dataset scripts to refer to it?
Hi ! You can host your text files for free on the Hugging Face Hub, just create a dataset repository and upload your files there. You can use the website to do so, or use git or python.