You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello,
first of all thanks for amazing research and for open sourcing the code & weights.
I had issues with training the TiTok & TA-TiTok with Single Stage loss, I tried:
training TiTok with the Single Stage loss
training TA-TiTok with Single Stage loss, but with placeholder text guidance prompt on ImageNet
In both cases I got very bad results: loss goes down, but the reconstructed image is just pale noise with not any real resemblence of the right colors or shapes. I trained with bs=32 on a singleA100 GPU for 10 hours. After which I would expect to start seeing signs of convergence. I noticed the grad norms are all in the 1e-7 to 1e-9 range.
Below are linked a minimal repro (~5 lines changed in config) and the resulting wandb training run with results.
Thank you for bringing this issue to our attention. We’ve identified that the problem was caused by the Perceptual Loss not being updated to align with the latest configuration. This has now been fixed in the latest update. We’ve verified that with the fix, the model begins reconstructing reasonable images around 25k steps with a total batch size of 256 on 8 A100 GPUs. Please give it a try and let us know if you encounter any further issues.
Hello,
first of all thanks for amazing research and for open sourcing the code & weights.
I had issues with training the TiTok & TA-TiTok with Single Stage loss, I tried:
In both cases I got very bad results: loss goes down, but the reconstructed image is just pale noise with not any real resemblence of the right colors or shapes. I trained with bs=32 on a single A100 GPU for 10 hours. After which I would expect to start seeing signs of convergence. I noticed the grad norms are all in the 1e-7 to 1e-9 range.
Below are linked a minimal repro (~5 lines changed in config) and the resulting wandb training run with results.
minimal diff repro
wandb training run report
The text was updated successfully, but these errors were encountered: