In order to run a model on GPU, two main changes are required:
1) When saving your TorchScript model, ensure that it is on the GPU.
For example, when using
pt2ts.py
,
this can be done by uncommenting the following lines:
device_type = torch.device("cuda")
trained_model = trained_model.to(device_type)
trained_model.eval()
trained_model_dummy_input_1 = trained_model_dummy_input_1.to(device_type)
trained_model_dummy_input_2 = trained_model_dummy_input_2.to(device_type)
Note: This code also moves the dummy input tensors to the GPU. Whilst not necessary for saving the model, but the tensors must also be on the GPU to test that the models runs.
2) When calling torch_tensor_from_array
in Fortran, the device type for the input
tensor(s) should be set to torch_kCUDA
, rather than torch_kCPU
.
This ensures that the inputs are on the same device type as the model.
Note: You do not need to change the device type for the output tensors as we want them to be on the CPU for subsequent use in Fortran.
In the case of having multiple GPU devices, as well as setting torch_kCUDA
as the
device type for any input tensors and models, you should also specify their device index
as the GPU device to be targeted. This argument is optional and will default to device
index 0 if unset.
For example, the following code snippet sets up a Torch tensor with GPU device index 2:
device_index = 2
call torch_tensor_from_array(in_tensors(1), in_data, tensor_layout, &
torch_kCUDA, device_index=device_index)
Whereas the following code snippet sets up a Torch tensor with (default) device index 0:
call torch_tensor_from_array(in_tensors(1), in_data, tensor_layout, &
torch_kCUDA)
See the MultiGPU example for a worked example of running with multiple GPUs.