You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Description level=error msg="rpc error: code = InvalidArgument desc = in ensemble 'deadlock_model', [request id: <id_unknown>] unexpected deadlock, at least one output is not set while no more ensemble steps can be made" func=blabla/grpc_proxy.server.callTritonProcedure file="/app/grpc_proxy/grpc_proxy_server.go:179
I sometimes met deadlock logs when getting requests and the frequency for this error is high when the server gets peak traffic. This log is from client that calls the triton model server in other container.
I add the config.pbtxt that is problematic and fixed, but the figure I added will be more intuitive.
Here's the problematic config.pbtxt that causes deadlock.
Here's the fixed config.pbtxt that doesn't cause deadlock. I just inferred that the main cause is using a model's output as the other models' input "and" the final output. So I just make A, B output not to directly go to final output but to pass C.
Triton Information
TRITON_VERSION=2.34.0 TRITON_CONTAINER_VERSION=23.05
I'm using custom build image that includes the backends only I need.
Expected behavior
I think the first config.pbtxt should work well, but it doesn't. I just want to know why the deadlock is caused.
The text was updated successfully, but these errors were encountered:
ukus04
changed the title
I think specific structure for ensemble model causes deadlock
Specific structure for ensemble model may causes deadlock
May 28, 2024
I'm encountering the same issue. It seems that system_shard_memory is being allocated using a combination of instance and name.I suspect that leading to a deadlock.
Description
level=error msg="rpc error: code = InvalidArgument desc = in ensemble 'deadlock_model', [request id: <id_unknown>] unexpected deadlock, at least one output is not set while no more ensemble steps can be made" func=blabla/grpc_proxy.server.callTritonProcedure file="/app/grpc_proxy/grpc_proxy_server.go:179
I sometimes met deadlock logs when getting requests and the frequency for this error is high when the server gets peak traffic. This log is from client that calls the triton model server in other container.
I add the config.pbtxt that is problematic and fixed, but the figure I added will be more intuitive.
Here's the problematic config.pbtxt that causes deadlock.
Here's the fixed config.pbtxt that doesn't cause deadlock. I just inferred that the main cause is using a model's output as the other models' input "and" the final output. So I just make A, B output not to directly go to final output but to pass C.
Triton Information
TRITON_VERSION=2.34.0 TRITON_CONTAINER_VERSION=23.05
I'm using custom build image that includes the backends only I need.
Expected behavior
I think the first config.pbtxt should work well, but it doesn't. I just want to know why the deadlock is caused.
The text was updated successfully, but these errors were encountered: