I have a server with 3 GPUs and I need to run inferences using all GPUs in my system to make full use of them, is there a code sample for it?
Hi @vilmara , which model do you want to use for inferences? Do you need to train it by yourself? Could you provide more detailed info about it. Thanks.
Hi @yhliang2018, it could be any pre-trained official model (preferably resnet50). I don't need to train it by myself, I just need a sample code that shows how to run inferences using efficiently all the GPUs in my system. Thanks
@vilmara TF hub would be a good start point if you want to use pre-trained models. It provides several tutorials to start with: https://www.tensorflow.org/hub/
As this is not an issue, I will close it for now.
@vilmara Were you able to find any code to run inferences on all GPUs efficiently in your system?
hi @SAswinGiridhar, have you explored Nvidia TensorRT Inference Server (TRTIS) ? https://github.com/NVIDIA/tensorrt-inference-server