| import torch |
|
|
|
|
| |
| def zluda_available(device_name): |
| return "[ZLUDA]" in device_name |
|
|
|
|
| |
| def enable_zluda_config(): |
| if hasattr(torch, 'cuda') and torch.cuda.is_available(): |
| device_name = torch.cuda.get_device_name(0) |
| print('Device name: ', device_name) |
| print('Cuda is available: ', torch.cuda.is_available()) |
| print('Cuda version: ', torch.version.cuda) |
| print('ZLUDA is available: ', zluda_available(device_name)) |
|
|
| if zluda_available(device_name): |
| torch.backends.cudnn.enabled = False |
| cuda_attr = torch.backends.cuda |
| if hasattr(cuda_attr, 'enable_flash_sdp'): |
| torch.backends.cuda.enable_flash_sdp(False) |
| print('Cuda enable flash sdp: ', False) |
| if hasattr(cuda_attr, 'enable_math_sdp'): |
| torch.backends.cuda.enable_math_sdp(True) |
| print('Cuda enable math sdp: ', True) |
| if hasattr(cuda_attr, 'enable_mem_efficient_sdp'): |
| torch.backends.cuda.enable_mem_efficient_sdp(False) |
| print('Cuda enable mem efficient sdp: ', False) |
| if hasattr(cuda_attr, 'enable_cudnn_sdp'): |
| torch.backends.cuda.enable_cudnn_sdp(False) |
| print('Cuda enable cudnn sdp: ', False) |
|
|