We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
TensorRT-LLM backend v0.20.0 release
https://github.com/triton-inference-server/tensorrtllm_backend/pull/748
TensorRT-LLM backend v0.18.2 release (#736)
TensorRT-LLM backend v0.18.1 release (#734)
TensorRT-LLM backend v0.18 release (#732)
update the submodule (#695)
TensorRT-LLM backend v0.16 release (#668)
TensorRT-LLM backend Release 0.15.0 (#654) Co-authored-by: Kaiyu Xie <26294424+kaiyux@users.noreply.github.com>
Update TensorRT-LLM backend v0.14.0 (#637)
TensorRT-LLM backend v0.13 Update (#607)