-
Hi everyone, I have a technical question regarding the fine-tuning process of the stable-diffusion-2-depth model, and I'd love to hear your insights. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 5 replies
-
Hi, the only people that can do a real Since the model is already supported by diffusers you just need to adapt the training script you want to use SD2, sadly that model architecture wasn't popular or really used so I don't think we have any training scripts for it. When you load the model with that |
Beta Was this translation helpful? Give feedback.
I think you're in the right track, don't really know if MM-DiT models are good for super resolution tasks, I don't use them normally because they're really slow and resource hungry, upscaling with them is even slower and you need a very high end GPU to be able to do it, and the benefits right now aren't that great to justify it.
What I do know is that the current SOTA model for super resolution is called SUPIR and it's based in SDXL, so a U-Net. Also there is a model called Stable Cascade that worked really well with a small latent and in the second stage upscale it with a second U-Net.
The other solution right now is to train or use a Tile ControlNet and do a tiled img2img over an image …