You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We are training the Stable Diffusion 3.5 Large model via Dreambooth method, using the sd3 branch.
Our system is an A100 Azure Remote Server.
Currently, we find that the only way to run inference on a trained model is via sd3_minimal_inference.py. We noticed that here are multiple manual operations done in this script including tokenizing and unpacking the .safetensors files of the model and text encoders.
Issues ⚠️
The only supported output format seems to be .safetensors - We compared the sd3_utils.py file to sdxl_utils.py and it seems that there is not yet support for the Diffusers format as there are many methods and conversions that seems to be absent.
Consecutively, there is not a compatible, easy way we found to load the 3 .safetensors files (model, clip_l, clip_g) to a pipe in order to run inference by ourselves, add sd3.5-compatible LoRAs on top, etc. We tried StableDiffusionPipeline, StableDiffusion3Pipeline and DiffusionPipeline all without luck of fixing the needed formats, dicts and metadata that is required in these libraries when loading the trained model.
Specifically regarding the text encoders, it seems that there is no metadata packed into the .safetensors file when saving. The config.json and such files which are normally saved into it and define the model size and parameters seem to be deliberately missing, we assume because in the sd3_minimal_inference.py file theses things are not needed and done manually. As such, we are unable to load them using libs such as CLIPModel from transformers and we are faced with incompatible format issues.
We are unable to use LoRAs, even via the sd3_minimal_inference.py script on our trained model. We try using the --lora_weights parameter (i.e tested LoRA). Is this related to the fact that 'Merging LoRAs from checkpoint' is regarded in the README as 'not yet supported'?
Thank you in advance to anyone replying and I apologize if anything aforementioned is trivial 🙏🏻
We are training the Stable Diffusion 3.5 Large model via Dreambooth method, using the sd3 branch.
Our system is an A100 Azure Remote Server.
Currently, we find that the only way to run inference on a trained model is via sd3_minimal_inference.py. We noticed that here are multiple manual operations done in this script including tokenizing and unpacking the .safetensors files of the model and text encoders.
Issues⚠️
The only supported output format seems to be .safetensors - We compared the sd3_utils.py file to sdxl_utils.py and it seems that there is not yet support for the Diffusers format as there are many methods and conversions that seems to be absent.
Consecutively, there is not a compatible, easy way we found to load the 3 .safetensors files (model, clip_l, clip_g) to a pipe in order to run inference by ourselves, add sd3.5-compatible LoRAs on top, etc. We tried StableDiffusionPipeline, StableDiffusion3Pipeline and DiffusionPipeline all without luck of fixing the needed formats, dicts and metadata that is required in these libraries when loading the trained model.
Specifically regarding the text encoders, it seems that there is no metadata packed into the .safetensors file when saving. The config.json and such files which are normally saved into it and define the model size and parameters seem to be deliberately missing, we assume because in the sd3_minimal_inference.py file theses things are not needed and done manually. As such, we are unable to load them using libs such as CLIPModel from transformers and we are faced with incompatible format issues.
We are unable to use LoRAs, even via the sd3_minimal_inference.py script on our trained model. We try using the --lora_weights parameter (i.e tested LoRA). Is this related to the fact that 'Merging LoRAs from checkpoint' is regarded in the README as 'not yet supported'?
Thank you in advance to anyone replying and I apologize if anything aforementioned is trivial 🙏🏻
Our Relevant Config 👨💻
The text was updated successfully, but these errors were encountered: