Skip to content

Conversation

@dxqb
Copy link
Collaborator

@dxqb dxqb commented Jan 7, 2026

When I increased the padding for offloaded tensors earlier, to fix this issue, #887, I have missed the following line:

+ len(tensors) * 4 # add enough padding for alignment

Should probably be 16 also. Apparently this didn't cause any issues so far, but it could be less efficient because the cache for activation offloading gets allocated too small?

Requires testing

When I increased the padding for offloaded tensors earlier, to fix this issue, #887, I have missed the following line:
https://github.com/Nerogar/OneTrainer/blob/5f6ed6a63effbc5afe10cf1162b869c0858279d2/modules/util/LayerOffloadConductor.py#L251

Apparently it didn't cause any issues, but it could be less efficient because the cache for activation offloading gets allocated too small?

Requires testing
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants