Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

docs: Fix broken links #7640

Closed
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion build.py
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@
#

# By default build.py builds the Triton Docker image, but can also be
# used to build without Docker. See docs/build.md and --help for more
# used to build without Docker. See docs/customization_guide/build.md and --help for more
# information.
#
# The TRITON_VERSION file indicates the Triton version and
Expand Down
2 changes: 1 addition & 1 deletion deploy/fleetcommand/values.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ image:
#
# To set model control mode, uncomment and configure below
# TODO: Fix the following url, it is invalid
# See https://github.com/triton-inference-server/server/blob/r24.08/docs/model_management.md
# See https://github.com/triton-inference-server/server/blob/r24.08/docs/user_guide/model_management.md
# for more details
#- --model-control-mode=explicit|poll|none
#
Expand Down
4 changes: 2 additions & 2 deletions docs/protocol/extension_statistics.md
Original file line number Diff line number Diff line change
Expand Up @@ -452,7 +452,7 @@ message InferStatistics
// model version, model inputs) hashes to an existing entry in the
// cache. On a cache miss, the request hash and response output tensor
// data is added to the cache. See response cache docs for more info:
// https://github.com/triton-inference-server/server/blob/main/docs/response_cache.md
// https://github.com/triton-inference-server/server/blob/main/docs/user_guide/response_cache.md
StatisticDuration cache_hit = 7;

// The count of response cache misses and cumulative duration to lookup
Expand All @@ -463,7 +463,7 @@ message InferStatistics
// miss occurs for a request to that model when the request metadata
// does NOT hash to an existing entry in the cache. See the response
// cache docs for more info:
// https://github.com/triton-inference-server/server/blob/main/docs/response_cache.md
// https://github.com/triton-inference-server/server/blob/main/docs/user_guide/response_cache.md
StatisticDuration cache_miss = 8;
}

Expand Down
8 changes: 4 additions & 4 deletions docs/user_guide/debugging_guide.md
Original file line number Diff line number Diff line change
Expand Up @@ -74,7 +74,7 @@ If your error message only occurs in one or a few places in the Triton code, you

**Step 3. Build with Debug Flags**

The next step is building with debug flags. We unfortunately don’t provide a debug container, so you’d need to follow the [build guide](https://github.com/triton-inference-server/server/blob/main/docs/customization_guide/build.md) to build the container, which includes a [section on adding debug symbols](https://github.com/triton-inference-server/server/blob/main/docs/build.md#building-with-debug-symbols). Once you do so, you can install GDB (`apt-get install gdb`) in the container and run Triton in GDB (`gdb --args tritonserver…`). If needed, you can open a second terminal to run a script in another container. If the server segfaults, you can enter `backtrace`, which will provide you a call stack that lets you know where the error got generated. You should then be able to trace the source of the error. If the bug still exists after debugging, we’ll need this to expedite our work.
The next step is building with debug flags. We unfortunately don’t provide a debug container, so you’d need to follow the [build guide](https://github.com/triton-inference-server/server/blob/main/docs/customization_guide/build.md) to build the container, which includes a [section on adding debug symbols](https://github.com/triton-inference-server/server/blob/main/docs/customization_guide/build.md#building-with-debug-symbols). Once you do so, you can install GDB (`apt-get install gdb`) in the container and run Triton in GDB (`gdb --args tritonserver…`). If needed, you can open a second terminal to run a script in another container. If the server segfaults, you can enter `backtrace`, which will provide you a call stack that lets you know where the error got generated. You should then be able to trace the source of the error. If the bug still exists after debugging, we’ll need this to expedite our work.

Advanced GDB users can also examine variable values, add breakpoints, and more to find the cause of their issue.

Expand All @@ -89,7 +89,7 @@ If you built the backend yourself, this could be a linking error. If you are con

## Server Issues

You generally should not run into errors with the server itself. If the server goes down, it’s usually because something went wrong during model loading or inference and you can use the above section to debug. It’s particularly useful to work through the [Building with Debug Flags](https://github.com/triton-inference-server/server/blob/main/docs/build.md#building-with-debug-symbols) section above to resolve those sorts of issues. However, this section will go through some specific cases that may occur.
You generally should not run into errors with the server itself. If the server goes down, it’s usually because something went wrong during model loading or inference and you can use the above section to debug. It’s particularly useful to work through the [Building with Debug Flags](https://github.com/triton-inference-server/server/blob/main/docs/customization_guide/build.md#building-with-debug-symbols) section above to resolve those sorts of issues. However, this section will go through some specific cases that may occur.

### No Connection to Server

Expand Down Expand Up @@ -121,9 +121,9 @@ We often get performance optimization questions around the clients. Triton clien

## Performance Issues

This section goes over debugging unexpected performance. If you are looking to optimize performance, please see the [Optimization](https://github.com/triton-inference-server/server/blob/main/docs/optimization.md) and [Performance Tuning](https://github.com/triton-inference-server/server/blob/main/docs/performance_tuning.md) guides.
This section goes over debugging unexpected performance. If you are looking to optimize performance, please see the [Optimization](https://github.com/triton-inference-server/server/blob/main/docs/user_guide/optimization.md) and [Performance Tuning](https://github.com/triton-inference-server/server/blob/main/docs/user_guide/performance_tuning.md) guides.

The easiest step to start with is running perf_analyzer to get a breakdown of the request lifecycle, throughput, and latency for each individual model. For a more detailed view, you can [enable tracing](https://github.com/triton-inference-server/server/blob/main/docs/trace.md) when running the server. This will provide exact timestamps to drill down into what is happening. You can also enable tracing with perf_analyzer for the GRPC and HTTP clients by using the tracing flags. Note that enabling tracing can impact Triton’s performance, but it can be helpful to examine the timestamps throughout a request’s lifecycle.
The easiest step to start with is running perf_analyzer to get a breakdown of the request lifecycle, throughput, and latency for each individual model. For a more detailed view, you can [enable tracing](https://github.com/triton-inference-server/server/blob/main/docs/user_guide/trace.md) when running the server. This will provide exact timestamps to drill down into what is happening. You can also enable tracing with perf_analyzer for the GRPC and HTTP clients by using the tracing flags. Note that enabling tracing can impact Triton’s performance, but it can be helpful to examine the timestamps throughout a request’s lifecycle.

### Performance Profiling

Expand Down
Loading