-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Issues: triton-inference-server/server
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Milestones
Assignee
Sort
Issues list
[ERROR] No available memory for the cache blocks.
#7562
opened Aug 22, 2024 by
TheNha
updated Jan 13, 2025
Server build with python BE failing due to missing Boost lib
#7925
opened Jan 9, 2025 by
buddhapuneeth
updated Jan 10, 2025
OpenAI-Compatible Frontend should support world_size larger than 1
enhancement
New feature or request
#7914
opened Jan 3, 2025 by
cocodee
updated Jan 9, 2025
Segfault/Coredump in grpc::ModelInferHandler::InferResponseComplete
crash
Related to server crashes, segfaults, etc.
grpc
Related to the GRPC server
#7877
opened Dec 12, 2024 by
andyblackheel
updated Jan 9, 2025
Support request cancellation on timeout for sync grpc client
enhancement
New feature or request
grpc
Related to the GRPC server
#7532
opened Aug 15, 2024 by
ShuaiShao93
updated Jan 9, 2025
Encountering a segmentation fault issue when attempting to send multiple images via gRPC
crash
Related to server crashes, segfaults, etc.
grpc
Related to the GRPC server
#6891
opened Feb 16, 2024 by
lawliet0823
updated Jan 9, 2025
Interaction of timeouts, ensemble scheduler and oldest sequence scheduler causes state leakage
bug
Something isn't working
grpc
Related to the GRPC server
#7117
opened Apr 15, 2024 by
jamied157
updated Jan 9, 2025
All gRPC requests to the Triton server are timing out, but HTTP requests are functioning normally.
grpc
Related to the GRPC server
investigating
The developement team is investigating this issue
#6899
opened Feb 21, 2024 by
SunnyGhj
updated Jan 9, 2025
Upgrading gRPC lib
enhancement
New feature or request
grpc
Related to the GRPC server
investigating
The developement team is investigating this issue
#7043
opened Mar 26, 2024 by
yutkin
updated Jan 9, 2025
grpc request performance issue
grpc
Related to the GRPC server
performance
A possible performance tune-up
#7187
opened May 7, 2024 by
Yuyaying-winnie
updated Jan 9, 2025
GRPC infer returns null in outputs contents
grpc
Related to the GRPC server
#7191
opened May 7, 2024 by
aohorodnyk
updated Jan 9, 2025
gRPC Segfaults in Triton 24.08 due to Low Request Cancellation Timeout
crash
Related to server crashes, segfaults, etc.
grpc
Related to the GRPC server
#7368
opened Jun 24, 2024 by
AshwinAmbal
updated Jan 9, 2025
Slow Dynamic Batching with High Concurrency on NVIDIA Triton Inference
grpc
Related to the GRPC server
module: server
Issues related to the server core
performance
A possible performance tune-up
#7492
opened Aug 2, 2024 by
tuanavu
updated Jan 9, 2025
GRPC Duplicate timer being added causing segfaults.
crash
Related to server crashes, segfaults, etc.
grpc
Related to the GRPC server
#7571
opened Aug 27, 2024 by
jamied157
updated Jan 9, 2025
50k-60k infer/sec limitation
grpc
Related to the GRPC server
module: server
Issues related to the server core
performance
A possible performance tune-up
#7590
opened Sep 3, 2024 by
v-hyhyniak-crt
updated Jan 9, 2025
Question : Clarification on Usage of passive Flag and Secondary Devices in Model Configurations
#7833
opened Nov 25, 2024 by
Hemaprasannakc
updated Jan 8, 2025
How to set the parameter make concurrent model execution?
performance
A possible performance tune-up
#7579
opened Aug 30, 2024 by
Will-Chou-5722
updated Jan 8, 2025
Specific structure for ensemble model may causes deadlock
#7280
opened May 28, 2024 by
ukus04
updated Jan 8, 2025
How to specify model version when platform: "ensemble"
#6188
opened Aug 15, 2023 by
efflerstegerd
updated Jan 7, 2025
How to get access to the vllm backend model
#7916
opened Jan 3, 2025 by
lianyiyi
updated Jan 3, 2025
running triton as a inference service on host
#7915
opened Jan 3, 2025 by
sriram-dsl
updated Jan 3, 2025
[Critical] Triton stops processing requests and crashes
bug
Something isn't working
#7649
opened Sep 24, 2024 by
appearancefnp
updated Jan 3, 2025
Python backend SHM memory leak
question
Further information is requested
#7481
opened Jul 27, 2024 by
mbahri
updated Jan 2, 2025
Failed to capture elapsed time: Internal - Failed to capture elapsed time: device not ready
bug
Something isn't working
#6700
opened Dec 14, 2023 by
zeruniverse
updated Jan 2, 2025
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.