Slackbot
01/20/2023, 12:43 AMlarme (shenyang)
01/20/2023, 5:29 AMtorch.cuda.is_available()
still return False
?larme (shenyang)
01/20/2023, 5:30 AMbentoml serve
, you should see vram usage in the output of nvidia-smi
, or else the model is not utilizing GPUlatemetal
01/20/2023, 4:56 PMlatemetal
01/20/2023, 4:56 PMlatemetal
01/20/2023, 4:56 PMlatemetal
01/20/2023, 4:57 PMPYTORCH_CUDA_ALLOC_CONF=max_split_size_mb:768
or lower otherwise i get memory errors