You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
At one time it leaked about 48GiB of memory, then I manually stopped it.
Reproduction steps
I run my image captioning script here with lots of parallel request (I tried 32): https://github.com/NeoChen1024/scripts/blob/master/llm-image-captioning.py
Then it will slowly leak memory, and also slowdown decoding stage (text generation t/s slowly declines, also the average GPU usage).
Enabling or disabling uvloop and CUDA malloc backend doesn't help.
Expected behavior
It doesn't leak memory, speed stays the same.
Logs
No response
Additional context
No response
Acknowledgements
I have looked for similar issues before submitting this one.
I have read the disclaimer, and this issue is related to a code bug. If I have a question, I will use the Discord server.
I understand that the developers have lives and my issue will be answered when possible.
I understand the developers of this program are human, and I will ask my questions politely.
The text was updated successfully, but these errors were encountered:
OS
Linux
GPU Library
CUDA 12.x
Python version
3.11
Describe the bug
When using Pixtral 12B VLM 8.0bpw (quantized by turboderp at https://huggingface.co/turboderp/pixtral-12b-exl2/tree/8.0bpw), the main process slowly leaks memory with lots of vision request.
At one time it leaked about 48GiB of memory, then I manually stopped it.
Reproduction steps
I run my image captioning script here with lots of parallel request (I tried 32): https://github.com/NeoChen1024/scripts/blob/master/llm-image-captioning.py
Then it will slowly leak memory, and also slowdown decoding stage (text generation t/s slowly declines, also the average GPU usage).
Enabling or disabling uvloop and CUDA malloc backend doesn't help.
Expected behavior
It doesn't leak memory, speed stays the same.
Logs
No response
Additional context
No response
Acknowledgements
The text was updated successfully, but these errors were encountered: