Torchserve server using a YoloV5 model running on docker with GPU and static batch inference to perform production ready and real time inference.
-
Updated
Feb 10, 2023 - Python
Torchserve server using a YoloV5 model running on docker with GPU and static batch inference to perform production ready and real time inference.
Serve pytorch inference requests using batching with redis for faster performance.
Add a description, image, and links to the batch-inference topic page so that developers can more easily learn about it.
To associate your repository with the batch-inference topic, visit your repo's landing page and select "manage topics."