Posted 2021-01-17Updated 2021-02-24MLOps3 minutes read (About 457 words)Parallelization of Inference on GPUs for TFServing using NginxA guiding to load balancing your TFServing Inference API over multiple GPUs.Read more