Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Quick questions on resources used for the full train #103

Open
glmcdona opened this issue Dec 23, 2024 · 0 comments
Open

Quick questions on resources used for the full train #103

glmcdona opened this issue Dec 23, 2024 · 0 comments

Comments

@glmcdona
Copy link

glmcdona commented Dec 23, 2024

Couple quick questions if possible! Looking to do a full finetune as an experiment with a i2v dataset I've been working on.

  • train.sh:
    • I note a 2e-5 learning rate. Was this the learning used for the full train of the model the entire time? Or was a learning rate schedule ran or something? Was 2e-5 the ending learning rate, or starting learning rate where it was lowered throughout the training? Trying to understand if this 2e-5 is a reasonable starting learning rate for a finetune, or if I should start it closer to 1e-4 as seen in the Lora config, then slowly ramp it down to 2e-5 towards the end.
    • train_batch_size of 1, does this mean a batch size of 1 was used? Should this be raised?
  • Training resources
    • How many iterations/steps were ran?
    • Curious approximately how much the full train cost and how long with what resources?
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant