-
Notifications
You must be signed in to change notification settings - Fork 449
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TensorRT #29
Comments
Hi @RuipingL, thanks for your interest in YOLO-World! We will provide the TensorRT version of YOLO-World shortly. Currently, could you try to deploy to TensorRT with the ONNX model? If you have any questions or make any progress, feel free to open issues. |
Hi @wondervictor, thanks for reply and happy Chinese New Year in advance. When transferring your model to ONNX format, an automatic dimension compression might occur. This can cause an error during the conversion process to TensorRT. |
Following this as well |
期待TensorRT版本,我有一个小疑问,请问预估能在Nvidia什么型号上的边缘设备进行25ms的推理呢?Orin Nx、Xavier Nx、Orin Nano?最后再次感谢你们精彩的工作! |
any updates on this issue? |
Maybe onnxruntime can use TensorRT ExecutionProvider. https://onnxruntime.ai/docs/execution-providers/TensorRT-ExecutionProvider.html |
https://github.com/PrinceP/tensorrt-cpp-for-onnx?tab=readme-ov-file#yolo-world Without dynamic batch, its working but onnx export has to be changed. |
Thanks for your awsome work. Could you release the TensorRT version of YOLO-World? The inference time of YOLO-World onnx is kind of slow.
The text was updated successfully, but these errors were encountered: