Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can unlimiformer work with common fine-tuning methods? #23

Open
mrlzh opened this issue Jul 18, 2023 · 1 comment
Open

Can unlimiformer work with common fine-tuning methods? #23

mrlzh opened this issue Jul 18, 2023 · 1 comment

Comments

@mrlzh
Copy link

mrlzh commented Jul 18, 2023

I set max_source_length with 10240 and training t5 models,but it ran out of CUDA memory.
I would like to know if unlimiformer can run together with fine-tuning methods such as LoRA.

@abertsch72
Copy link
Owner

Hi @mrlzh , thanks for your interest!

We haven't tried using Unlimiformer with LoRA, but there isn't a theoretical reason that they wouldn't work together. If you try it, please let us know how it goes!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants