You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: PyTorch/SpeechSynthesis/Tacotron2/trt/README.md
+26-32Lines changed: 26 additions & 32 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,19 +1,11 @@
1
-
# Tacotron 2 and WaveGlow Inference For TensorRT
2
-
3
-
This is subfolder of the Tacotron 2 for PyTorch repository, tested and
4
-
maintained by NVIDIA, and provides scripts to perform high-performance
5
-
inference using NVIDIA TensorRT.
6
-
The Tacotron 2 and WaveGlow models form a text-to-speech (TTS) system that
7
-
enables users to synthesize natural sounding speech from raw transcripts
8
-
without any additional information such as patterns and/or rhythms of speech.
9
-
More information about the TTS system and its training can be found in the
1
+
# Tacotron 2 and WaveGlow Inference with TensorRT
2
+
3
+
This is subfolder of the Tacotron 2 for PyTorch repository, tested and maintained by NVIDIA, and provides scripts to perform high-performance inference using NVIDIA TensorRT.
4
+
5
+
The Tacotron 2 and WaveGlow models form a text-to-speech (TTS) system that enables users to synthesize natural sounding speech from raw transcripts without any additional information such as patterns and/or rhythms of speech. More information about the TTS system and its training can be found in the
10
6
[Tacotron 2 PyTorch README](../README.md).
11
-
NVIDIA TensorRT is a platform for high-performance deep learning inference.
12
-
It includes a deep learning inference optimizer and runtime that delivers low
13
-
latency and high-throughput for deep learning inference applications. After
14
-
optimizing the compute-intensive acoustic model with NVIDIA TensorRT,
15
-
inference throughput increased by up to 1.4x over native PyTorch in mixed
16
-
precision.
7
+
8
+
NVIDIA TensorRT is a platform for high-performance deep learning inference. It includes a deep learning inference optimizer and runtime that delivers low latency and high-throughput for deep learning inference applications. After optimizing the compute-intensive acoustic model with NVIDIA TensorRT, inference throughput increased by up to 1.4x over native PyTorch in mixed precision.
3. Build the Tacotron 2 and WaveGlow PyTorch NGC container.
@@ -49,10 +40,18 @@ and store them in `./checkpoints` directory:
49
40
bash scripts/docker/interactive.sh
50
41
```
51
42
52
-
5. Export the models to ONNX intermediate representations (ONNX IRs).
43
+
5. Verify that TensorRT version installed is 7.0 or greater. If necessary, download and install the latest release from https://developer.nvidia.com/nvidia-tensorrt-download
44
+
45
+
```bash
46
+
pip list | grep tensorrt
47
+
dpkg -l | grep TensorRT
48
+
```
49
+
50
+
6. Export the models to ONNX intermediate representation (ONNX IR).
53
51
Export Tacotron 2 to three ONNX parts: Encoder, Decoder, and Postnet:
Our results were obtained by running the `./trt/run_latency_tests_trt.sh` script in
87
-
the PyTorch-19.11-py3 NGC container. Please note that to reproduce the results,
88
-
you need to provide pretrained checkpoints for Tacotron 2 and WaveGlow. Please
89
-
edit the script to provide your checkpoint filenames. For all tests in this table,
90
-
we used WaveGlow with 256 residual channels.
84
+
Our results were obtained by running the `./trt/run_latency_tests_trt.sh` script in the PyTorch-19.11-py3 NGC container. Please note that to reproduce the results, you need to provide pretrained checkpoints for Tacotron 2 and WaveGlow. Please edit the script to provide your checkpoint filenames. For all tests in this table, we used WaveGlow with 256 residual channels.
0 commit comments