Triton-rust is a gRPC library to interact with Nvidia Triton Inference Server.
You can compile the library together with all examples with:
make all
You can find several examples of neural network inference using Triton Inference Server and Rust. These examples could be found here.
- CUDA shared memory is not functionnal yet
- Rust's ndarrays are to be in standard layout
Boris Albar (b.albar@catie.fr)
This work has been done in frame of the Vaniila platform.