Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implementing support for PIXART-α Fine Tuning and DreamBooth #979

Open
FurkanGozukara opened this issue Dec 3, 2023 · 4 comments
Open

Comments

@FurkanGozukara
Copy link

PIXART-α: Fast Training of Diffusion Transformer for Photorealistic Text-to-Image Synthesis

https://pixart-alpha.github.io/

They have train scripts here : https://github.com/PixArt-alpha/PixArt-alpha/tree/master/train_scripts

This model is literally better than SDXL

I made a full tutorial for those who wonders.

PIXART-α : First Open Source Rival to Midjourney - Better Than Stable Diffusion SDXL - Full Tutorial

image

@mrT333
Copy link

mrT333 commented Dec 4, 2023

This model is literally better than SDXL

And what is most amazing about it: It matches SDXL quality while having only 12% training time of SD v1.5 (remember the low quality of base SD v1.5?). This could indicate that once the community starts finetuning Pixart it will result in much higher quality improvements using way less training steps than SD 1.5 or SDXL.

Also we might not even know where the quality ceiling of Pixart is as it could be severely undertrained with just 12% training steps... meaning finetuning would have even greater impact on quality than it did on SD 1.5 and it could reach a quality plateau way later than all stable diffusion based models.

@FurkanGozukara
Copy link
Author

This model is literally better than SDXL

And what is most amazing about it: It matches SDXL quality while having only 12% training time of SD v1.5 (remember the low quality of base SD v1.5?). This could indicate that once the community starts finetuning Pixart it will result in much higher quality improvements using way less training steps than SD 1.5 or SDXL.

Also we might not even know where the quality ceiling of Pixart is as it could be severely undertrained with just 12% training steps... meaning finetuning would have even greater impact on quality than it did on SD 1.5 and it could reach a quality plateau way later than all stable diffusion based models.

this is 100% accurate. huge potential there is.

@storuky
Copy link

storuky commented Dec 10, 2023

This model is literally better than SDXL

It matches SDXL quality while having only 12% training time of SD v1.5

Lol, this model is totally undertrained. Just type something like "a giraffe" and generate few results with different seeds. Same with tiger, lion, etc. It's not variative.

Once again I'm convinced that if someone talks about reducing CO2 emissions, it's empty talk.

@FurkanGozukara
Copy link
Author

CO2 emissions

who cares about CO2 emissions :)

we need better model

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants