Skip to content
This repository was archived by the owner on Aug 7, 2025. It is now read-only.

Commit 781cb51

Browse files
authored
Merge branch 'master' into fix/dali_batch_input
2 parents 6dd7b66 + d31b6c3 commit 781cb51

File tree

4 files changed

+5
-102
lines changed

4 files changed

+5
-102
lines changed

docs/conf.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@
1111
#
1212
# All configuration values have a default; values that are commented out
1313
# serve to show the default.
14-
14+
#
1515
# If extensions (or modules to document with autodoc) are in another directory,
1616
# add these directories to sys.path here. If the directory is relative to the
1717
# documentation root, use os.path.abspath to make it absolute, like shown here.
@@ -131,6 +131,7 @@
131131
"collapse_navigation": True,
132132
"display_version": True,
133133
"logo_only": True,
134+
"analytics_id": "GTM-T8XT4PS",
134135
}
135136

136137
html_logo = "_static/img/pytorch-logo-dark.svg"
@@ -239,7 +240,6 @@ def setup(app):
239240

240241
# Register custom directives
241242

242-
243243
rst.directives.register_directive("devices", SupportedDevices)
244244
rst.directives.register_directive("properties", SupportedProperties)
245245
rst.directives.register_directive("customcardstart", CustomCardStart)

docs/indexOLD.md

Lines changed: 0 additions & 98 deletions
This file was deleted.

docs/large_model_inference.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ In addition to this default behavior, TorchServe provides the flexibility for us
1212

1313
Using Pippy integration as an example, the image below illustrates the internals of the TorchServe large model inference.
1414

15-
![ts-lmi-internal](images/ts-lmi-internal.png)
15+
![ts-lmi-internal](https://raw.githubusercontent.com/pytorch/serve/master/docs/images/ts-lmi-internal.png)
1616

1717
## PiPPy (PyTorch Native solution for large model inference)
1818

@@ -186,7 +186,7 @@ torch-model-archiver --model-name bloom --version 1.0 --handler deepspeed_handle
186186
#### Tune "[responseTimeout](https://github.com/pytorch/serve/blob/5ee02e4f050c9b349025d87405b246e970ee710b/docs/configuration.md?plain=1#L216)" (see [model config YAML file](https://github.com/pytorch/serve/blob/5ee02e4f050c9b349025d87405b246e970ee710b/model-archiver/README.md?plain=1#L164)) if high model loading or inference latency causes response timeout.
187187

188188
#### Tune torchrun parameters
189-
User is able to tune torchrun parameters in [model config YAML file](https://github.com/pytorch/serve/blob/2f1f52f553e83703b5c380c2570a36708ee5cafa/model-archiver/README.md?plain=1#L179). The supported parameters are defined at [here](https://github.com/pytorch/serve/blob/2f1f52f553e83703b5c380c2570a36708ee5cafa/frontend/archive/src/main/java/org/pytorch/serve/archive/model/ModelConfig.java#L329). For example, by default, `OMP_NUMNER_T?HREADS` is 1. It can be modified in the YAML file.
189+
User is able to tune torchrun parameters in [model config YAML file](https://github.com/pytorch/serve/blob/2f1f52f553e83703b5c380c2570a36708ee5cafa/model-archiver/README.md?plain=1#L179). The supported parameters are defined at [here](https://github.com/pytorch/serve/blob/2f1f52f553e83703b5c380c2570a36708ee5cafa/frontend/archive/src/main/java/org/pytorch/serve/archive/model/ModelConfig.java#L329). For example, by default, `OMP_NUMBER_THREADS` is 1. It can be modified in the YAML file.
190190
```yaml
191191
#frontend settings
192192
torchrun:

ts_scripts/spellcheck_conf/wordlist.txt

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1062,3 +1062,4 @@ XLA
10621062
inferentia
10631063
ActionSLAM
10641064
statins
1065+
chatGPT

0 commit comments

Comments
 (0)