Skip to content
This repository was archived by the owner on Aug 7, 2025. It is now read-only.

Commit d31b6c3

Browse files
authored
G analytics (#2449)
* Adding Google Analytics Adding Google Analytics to TorchServe pages for metrics. Removing indexOld.md file * Update large_model_inference.md * Spellcheck update Adding chatGPT to wordslist.txt * Update conf.py * Update conf.py * Update conf.py
1 parent 4d7dc64 commit d31b6c3

File tree

4 files changed

+5
-102
lines changed

4 files changed

+5
-102
lines changed

docs/conf.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@
1111
#
1212
# All configuration values have a default; values that are commented out
1313
# serve to show the default.
14-
14+
#
1515
# If extensions (or modules to document with autodoc) are in another directory,
1616
# add these directories to sys.path here. If the directory is relative to the
1717
# documentation root, use os.path.abspath to make it absolute, like shown here.
@@ -131,6 +131,7 @@
131131
"collapse_navigation": True,
132132
"display_version": True,
133133
"logo_only": True,
134+
"analytics_id": "GTM-T8XT4PS",
134135
}
135136

136137
html_logo = "_static/img/pytorch-logo-dark.svg"
@@ -239,7 +240,6 @@ def setup(app):
239240

240241
# Register custom directives
241242

242-
243243
rst.directives.register_directive("devices", SupportedDevices)
244244
rst.directives.register_directive("properties", SupportedProperties)
245245
rst.directives.register_directive("customcardstart", CustomCardStart)

docs/indexOLD.md

Lines changed: 0 additions & 98 deletions
This file was deleted.

docs/large_model_inference.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ In addition to this default behavior, TorchServe provides the flexibility for us
1212

1313
Using Pippy integration as an example, the image below illustrates the internals of the TorchServe large model inference.
1414

15-
![ts-lmi-internal](images/ts-lmi-internal.png)
15+
![ts-lmi-internal](https://raw.githubusercontent.com/pytorch/serve/master/docs/images/ts-lmi-internal.png)
1616

1717
## PiPPy (PyTorch Native solution for large model inference)
1818

@@ -186,7 +186,7 @@ torch-model-archiver --model-name bloom --version 1.0 --handler deepspeed_handle
186186
#### Tune "[responseTimeout](https://github.com/pytorch/serve/blob/5ee02e4f050c9b349025d87405b246e970ee710b/docs/configuration.md?plain=1#L216)" (see [model config YAML file](https://github.com/pytorch/serve/blob/5ee02e4f050c9b349025d87405b246e970ee710b/model-archiver/README.md?plain=1#L164)) if high model loading or inference latency causes response timeout.
187187

188188
#### Tune torchrun parameters
189-
User is able to tune torchrun parameters in [model config YAML file](https://github.com/pytorch/serve/blob/2f1f52f553e83703b5c380c2570a36708ee5cafa/model-archiver/README.md?plain=1#L179). The supported parameters are defined at [here](https://github.com/pytorch/serve/blob/2f1f52f553e83703b5c380c2570a36708ee5cafa/frontend/archive/src/main/java/org/pytorch/serve/archive/model/ModelConfig.java#L329). For example, by default, `OMP_NUMNER_T?HREADS` is 1. It can be modified in the YAML file.
189+
User is able to tune torchrun parameters in [model config YAML file](https://github.com/pytorch/serve/blob/2f1f52f553e83703b5c380c2570a36708ee5cafa/model-archiver/README.md?plain=1#L179). The supported parameters are defined at [here](https://github.com/pytorch/serve/blob/2f1f52f553e83703b5c380c2570a36708ee5cafa/frontend/archive/src/main/java/org/pytorch/serve/archive/model/ModelConfig.java#L329). For example, by default, `OMP_NUMBER_THREADS` is 1. It can be modified in the YAML file.
190190
```yaml
191191
#frontend settings
192192
torchrun:

ts_scripts/spellcheck_conf/wordlist.txt

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1062,3 +1062,4 @@ XLA
10621062
inferentia
10631063
ActionSLAM
10641064
statins
1065+
chatGPT

0 commit comments

Comments
 (0)