Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

instance recommendation for full graph #7

Open
laurendonoghue opened this issue Feb 5, 2025 · 0 comments
Open

instance recommendation for full graph #7

laurendonoghue opened this issue Feb 5, 2025 · 0 comments

Comments

@laurendonoghue
Copy link

Congratulations on this exciting new method!

Could you please provide clarity on what kind of GPU resources are recommended to train with the full graph? As kgwas.py train() does not allow for the specification of multiple CUDA devices, does that suggest this should be able to run on just 1 device?

I am continually running into CUDA OOM errors even when lowering the batch size significantly. I can train through 10 epochs (memory reaching near 21GB out of 24 GB available), however, for reasons unclear to me it will error out during the final step of saving the model predictions/results. If relevant, I also had to change num_workers=1 to avoid DataLoader errors.

I have been able to train through 10 epochs/save results with the full graph on CPU high memory, so I know I am not encountering other issues, but that is of course undesirably slow.

Additionally, given that all of the benchmarking done in the paper was done with the full graph, did you do any benchmarking between the "fast" Enformer/ESM graph and the full graph?

Thanks for your guidance!

@laurendonoghue laurendonoghue changed the title instance recommendation for using full graph instance recommendation for full graph Feb 6, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant