Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How are the TiTok VAE models trained? #72

Open
Jyouhou opened this issue Jan 17, 2025 · 1 comment
Open

How are the TiTok VAE models trained? #72

Jyouhou opened this issue Jan 17, 2025 · 1 comment

Comments

@Jyouhou
Copy link

Jyouhou commented Jan 17, 2025

Great work!

Interested in the details of training the VAE variants.

(1) Are they also trained via the 2-stage procedure?
(2) what losses are used?

@cornettoyu
Copy link
Collaborator

Hi,

The new checkpoints (VAE or VQ) are trained with single-stage recipe. Losses include L2, perceptual loss (lpips + convnext), and gan loss. You can find more details in MaskBit and MaskGen tech report. We expect the codebase to be updated with the updated singe-stage training recipe in the near future and please stay tuned :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants