Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

how to train the pretrained Latte model weights? #256

Open
leonardodora opened this issue Apr 19, 2024 · 3 comments
Open

how to train the pretrained Latte model weights? #256

leonardodora opened this issue Apr 19, 2024 · 3 comments

Comments

@leonardodora
Copy link

as the vae of opensora is different from Latte, the weights from latte could be able to use directly?
Or your team train a latte model from scratch?

@LinB203
Copy link
Member

LinB203 commented Apr 20, 2024

Yes, we use it directly. It adapts very quickly and the transformation is visible in about 500 steps. This is consistent with pixart-sigma.

@leonardodora
Copy link
Author

Yes, we use it directly. It adapts very quickly and the transformation is visible in about 500 steps. This is consistent with pixart-sigma.

Thanks for your reply! But if we want to scale up the param number, what should we do first, you suggest? just train a larger new latte model?

@LinB203
Copy link
Member

LinB203 commented Apr 25, 2024

Yes, we use it directly. It adapts very quickly and the transformation is visible in about 500 steps. This is consistent with pixart-sigma.

Thanks for your reply! But if we want to scale up the param number, what should we do first, you suggest? just train a larger new latte model?

I think a pixart-alpha should be retrained.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants