Why is pytorch cuda total memory not aligned with the memory size of GPU cluster I created?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
02-01-2023 08:22 AM
No matter GPU cluster of which size I create, cuda total capacity is always ~16 Gb. Does anyone know what is the issue?
The code I use to get the total capacity:
torch.cuda.get_device_properties(0).total_memory
- Labels:
-
Memory Size
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
02-01-2023 10:11 PM
Hi, Could you please let us know which DBR has been installed? Also, please let us know if you have gone the supported instance types. Reference: https://docs.databricks.com/clusters/gpu.html.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
02-02-2023 12:46 AM
Hi, the DBR version is 11.3 LTS ML. The instance types I created is g4dn.

- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
04-08-2023 07:30 PM
Hi @Simon Zhang
Hope everything is going great.
Just wanted to check in if you were able to resolve your issue. If yes, would you be happy to mark an answer as best so that other members can find the solution more quickly? If not, please tell us so we can help you.
Cheers!

