Replies: 1 comment 1 reply
-
I think it worth a try. You can simply use the script here https://github.com/OpenLMLab/collie/blob/dev/examples/profile/memory_optim.py, with just modify the 2048 to 4096 :) |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I was wondering if it is possible to train on 4 A6000s with the 4096 context length option.
I think it's probably not possible, but I ask just in case.
Beta Was this translation helpful? Give feedback.
All reactions