Skip to content

Issues: triton-inference-server/server

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[ERROR] No available memory for the cache blocks.
#7562 opened Aug 22, 2024 by TheNha updated Jan 13, 2025
Server build with python BE failing due to missing Boost lib
#7925 opened Jan 9, 2025 by buddhapuneeth updated Jan 10, 2025
Triton crashes with SIGSEGV (signal 11) crash Related to server crashes, segfaults, etc. grpc Related to the GRPC server question Further information is requested
#7472 opened Jul 24, 2024 by JindrichD updated Jan 10, 2025
OpenAI-Compatible Frontend should support world_size larger than 1 enhancement New feature or request
#7914 opened Jan 3, 2025 by cocodee updated Jan 9, 2025
Segfault/Coredump in grpc::ModelInferHandler::InferResponseComplete crash Related to server crashes, segfaults, etc. grpc Related to the GRPC server
#7877 opened Dec 12, 2024 by andyblackheel updated Jan 9, 2025
Support request cancellation on timeout for sync grpc client enhancement New feature or request grpc Related to the GRPC server
#7532 opened Aug 15, 2024 by ShuaiShao93 updated Jan 9, 2025
Encountering a segmentation fault issue when attempting to send multiple images via gRPC crash Related to server crashes, segfaults, etc. grpc Related to the GRPC server
#6891 opened Feb 16, 2024 by lawliet0823 updated Jan 9, 2025
Interaction of timeouts, ensemble scheduler and oldest sequence scheduler causes state leakage bug Something isn't working grpc Related to the GRPC server
#7117 opened Apr 15, 2024 by jamied157 updated Jan 9, 2025
All gRPC requests to the Triton server are timing out, but HTTP requests are functioning normally. grpc Related to the GRPC server investigating The developement team is investigating this issue
#6899 opened Feb 21, 2024 by SunnyGhj updated Jan 9, 2025
Upgrading gRPC lib enhancement New feature or request grpc Related to the GRPC server investigating The developement team is investigating this issue
#7043 opened Mar 26, 2024 by yutkin updated Jan 9, 2025
grpc request performance issue grpc Related to the GRPC server performance A possible performance tune-up
#7187 opened May 7, 2024 by Yuyaying-winnie updated Jan 9, 2025
GRPC infer returns null in outputs contents grpc Related to the GRPC server
#7191 opened May 7, 2024 by aohorodnyk updated Jan 9, 2025
gRPC Segfaults in Triton 24.08 due to Low Request Cancellation Timeout crash Related to server crashes, segfaults, etc. grpc Related to the GRPC server
#7368 opened Jun 24, 2024 by AshwinAmbal updated Jan 9, 2025
Slow Dynamic Batching with High Concurrency on NVIDIA Triton Inference grpc Related to the GRPC server module: server Issues related to the server core performance A possible performance tune-up
#7492 opened Aug 2, 2024 by tuanavu updated Jan 9, 2025
GRPC Duplicate timer being added causing segfaults. crash Related to server crashes, segfaults, etc. grpc Related to the GRPC server
#7571 opened Aug 27, 2024 by jamied157 updated Jan 9, 2025
50k-60k infer/sec limitation grpc Related to the GRPC server module: server Issues related to the server core performance A possible performance tune-up
#7590 opened Sep 3, 2024 by v-hyhyniak-crt updated Jan 9, 2025
How to set the parameter make concurrent model execution? performance A possible performance tune-up
#7579 opened Aug 30, 2024 by Will-Chou-5722 updated Jan 8, 2025
Specific structure for ensemble model may causes deadlock
#7280 opened May 28, 2024 by ukus04 updated Jan 8, 2025
How to specify model version when platform: "ensemble"
#6188 opened Aug 15, 2023 by efflerstegerd updated Jan 7, 2025
How to get access to the vllm backend model
#7916 opened Jan 3, 2025 by lianyiyi updated Jan 3, 2025
running triton as a inference service on host
#7915 opened Jan 3, 2025 by sriram-dsl updated Jan 3, 2025
[Critical] Triton stops processing requests and crashes bug Something isn't working
#7649 opened Sep 24, 2024 by appearancefnp updated Jan 3, 2025
Python backend SHM memory leak question Further information is requested
#7481 opened Jul 27, 2024 by mbahri updated Jan 2, 2025
Failed to capture elapsed time: Internal - Failed to capture elapsed time: device not ready bug Something isn't working
#6700 opened Dec 14, 2023 by zeruniverse updated Jan 2, 2025
ProTip! Exclude everything labeled bug with -label:bug.