I have started creating a synthetic dataset consisting of the original images and their segmentation masks (one black/white mask as a separate png image for each instance in the original image). My aim is to use my dataset for finetuning on Mask2Former.
As far as I understood according to this link, I need to merge all masks into one segmentation map for each original image. Is that right? If so, this shouldn’t be a problem.
For this step I don’t want to use segments.ai as suggested in this blog, because I created my dataset with Blender and it’s quite simple to create segmentation maps from there with a python script.
So now I have original images and one segmentation map for each image.
What are the next steps for me? I somehow have to link the RGB colors of the masks to class names (in a JSON file? which format?), but how exactly does this need to look like?
And what else do I need to do to create a hugging face dataset thats suited for finetuning Mask2Former?
As I am quite new to the world of AI/datasets/huggingface I might need you to explain it in small steps, even if the problem seems easy for you to solve . Any help is appreciated. Thanks a lot!