Support for Nvidia Tesla M40 cards? [update] Can work. Don't do it. #90
Replies: 4 comments 6 replies
-
Unrelated but also of note. I got this error when running ./download_models.sh After inspecting the logs it turns out the container was running out of memory. I had the VM it is in set to 4GB. Changing that to 8GB allowed the process to complete properly. |
Beta Was this translation helpful? Give feedback.
-
Did you buy the Tesla M40 specifically for WIS? We only recommend Pascal (device capability 6) series GPUs and up which for a Tesla would be a P4 (P = Pascal). Note that a Tesla is not a Tesla - it's the name of a product line and there are drastic differences between M40, P4, T4, etc. The Tesla M40 is Maxwell (almost 10 years old) and we have no plans to officially support it. If anything we'll probably add a check for minimum compute 60 (we drop the decimal internally) and disable CUDA if less. As you note it doesn't support int8/float16 and it's more-or-less end of support - Maxwell is dropped in CUDA 12 (which we will be switching to shortly). It's also just a terrible card all around - dual slot, 250 watts(!!!), horrible performance. The P4 came out roughly a year later, is the same price or less, and is night and day better in every way (VRAM on any of these cards isn't close to an issue). In short, if you bought it for WIS you should return it (hopefully you can) and get a P4. |
Beta Was this translation helpful? Give feedback.
-
FYI, I got the M40 to work with Willow pretty easily, I just changed |
Beta Was this translation helpful? Give feedback.
-
As a compromise (for now) we'll detect pre-pascal cards, set float32, and heavily warn |
Beta Was this translation helpful? Give feedback.
-
I had to modify main.py under the section "# Override compute_type if at least one non-Turing card" to change compute_type to float32 otherwise I was unable to get the model running.
Apparently the Tesla M40 does not support int8 and since its cuda_device_capability is '52' and below '70', setting to int8 keeps the system from loading.
I was able to get everything up and running but have yet to test it properly or to get willow pointed to it, I will update with results later.
I found this chart with the supported compute types. I wish I had known before I went out and bout the Tesla card or I would have bought one of the other cards on the list. Maybe this will help someone else as well.
https://docs.nvidia.com/deeplearning/tensorrt/support-matrix/index.html#hardware-precision-matrix
Beta Was this translation helpful? Give feedback.
All reactions