Any advice on how to make it work would be most welcome.
Thanks very much
EDIT:
I’ve tried loading the SDXL base from the StableDiffusionXLImg2ImgPipeline but the results are strange, it’s returning the same image but worse, its seeing something and trying to process it but it’s not nearly the same as SD1.5.
I’m not sure where I’m going wrong, please help peeps.
init_image is the original image used in the base model, has to be exact size of the output: init_image = PIL.Image.open(init_image_path).convert(“RGB”)
the base model needs to return latents - not PIL.Image.
the ‘latents’ parameter in the function for the refiner has to be the .images of the return object of the base model (typeof diffusers.pipelines.pipeline_utils.ImagePipelineOutput)
For me the example in diffusers\pipelines\stable_diffusion_xl\pipeline_stable_diffusion_xl_img2img.py (line 52) helped me a lot…
I’m getting great results with good images to start, when I use finger drawn images I’m getting terrible results, the idea is to take the simple drawing and make it real. I get great results with the SD1.5 img to img and I just cant figure this one out. Please see my images as an example.
I’m resizing all images to 1024x1024 for best results