-
Notifications
You must be signed in to change notification settings - Fork 53
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Extremely slow training speed when using multiple GPU #23
Comments
I seldom use DDP and have tested on several simple scenes but did not meet the problem you mentioned. So did not have any ideas about it when you opened this issue. |
meet same issue. |
|
yeah, the situation i mentioned above is about the new strategy. |
Would you mind providing the config file content and command you used to run the training? |
I modified mcmc_density_controller so that it can work in distributed training. |
I do not have ideas about yours either. Maybe you can enable profiler with the option |
Hi, thanks for open-sourcing such great work!
The introduced multi-gpu parallel training is indeed useful. However, I found it extremely slow. Take the rubble dataset of Mill19 as an example, I trained a 3DGS with 60000 iterations. The first half training with densification on single A100 costs 0.8 hours, but the later half training on 4 A100 costs 10.1 hours. Here is my script, and large_scale.yaml only sets optimization parameters.
I'm wondering if you have met similar problems or have any ideas about solutions. Thanks!
The text was updated successfully, but these errors were encountered: