|
1 | 1 | # TorchServe
|
2 | 2 |
|
3 |
| -TorchServe is a flexible and easy to use tool for serving PyTorch models. |
| 3 | +TorchServe is a performant, flexible and easy to use tool for serving PyTorch eager mode and torschripted models. |
4 | 4 |
|
5 | 5 | ## Basic Features
|
6 | 6 |
|
7 | 7 | * [Serving Quick Start](https://github.com/pytorch/serve/blob/master/README.md#serve-a-model) - Basic server usage tutorial
|
8 | 8 | * [Model Archive Quick Start](https://github.com/pytorch/serve/tree/master/model-archiver#creating-a-model-archive) - Tutorial that shows you how to package a model archive file.
|
9 | 9 | * [Installation](https://github.com/pytorch/serve/blob/master/README.md#install-torchserve) - Installation procedures
|
10 |
| -* [Serving Models](server.md) - Explains how to use torchserve |
11 |
| - * [REST API](rest_api.md) - Specification on the API endpoint for TorchServe |
| 10 | +* [Serving Models](server.md) - Explains how to use TorchServe |
| 11 | +* [REST API](rest_api.md) - Specification on the API endpoint for TorchServe |
| 12 | +* [gRPC API](grpc_api.md) - TorchServe supports gRPC APIs for both inference and management calls |
12 | 13 | * [Packaging Model Archive](https://github.com/pytorch/serve/tree/master/model-archiver#torch-model-archiver-for-torchserve) - Explains how to package model archive file, use `model-archiver`.
|
13 | 14 | * [Inference API](inference_api.md) - How to check for the health of a deployed model and get inferences
|
14 | 15 | * [Management API](management_api.md) - How to manage and scale models
|
15 | 16 | * [Logging](logging.md) - How to configure logging
|
16 | 17 | * [Metrics](metrics.md) - How to configure metrics
|
17 |
| - * [Metrics API](metrics_api.md) - How to configure metrics API |
| 18 | +* [Prometheus and Grafana metrics](metrics_api.md) - How to configure metrics API with Prometheus formatted metrics in a Grafana dashboard |
| 19 | +* [Captum Explanations](https://github.com/pytorch/serve/blob/master/captum/Captum_visualization_for_bert.ipynb) - Built in support for Captum explanations for both text and images |
18 | 20 | * [Batch inference with TorchServe](batch_inference_with_ts.md) - How to create and serve a model with batch inference in TorchServe
|
19 | 21 | * [Workflows](workflows.md) - How to create workflows to compose Pytorch models and Python functions in sequential and parallel pipelines
|
20 |
| -* [Model Zoo](model_zoo.md) - List of pre-trained model archives ready to be served for inference with TorchServe. |
21 |
| -* [Examples](https://github.com/pytorch/serve/tree/master/examples) - Many examples of how to package and deploy models and workflows with TorchServe |
22 | 22 |
|
23 |
| -## Advanced Features |
24 | 23 |
|
25 |
| -* [Advanced configuration](configuration.md) - Describes advanced TorchServe configurations. |
26 |
| -* [Custom Service](custom_service.md) - Describes how to develop custom inference services. |
27 |
| -* [Unit Tests](https://github.com/pytorch/serve/tree/master/ts/tests#testing-torchserve) - Housekeeping unit tests for TorchServe. |
28 |
| -* [Benchmark](https://github.com/pytorch/serve/tree/master/benchmarks#torchserve-model-server-benchmarking) - Use JMeter to run TorchServe through the paces and collect benchmark data. |
29 |
| -* [TorchServe on Kubernetes](https://github.com/pytorch/serve/blob/master/kubernetes/README.md#torchserve-on-kubernetes) - Demonstrates a Torchserve deployment in Kubernetes using Helm Chart. |
30 | 24 |
|
31 | 25 | ## Default Handlers
|
32 | 26 |
|
33 | 27 | * [Image Classifier](https://github.com/pytorch/serve/blob/master/ts/torch_handler/image_classifier.py) - This handler takes an image and returns the name of object in that image
|
34 | 28 | * [Text Classifier](https://github.com/pytorch/serve/blob/master/ts/torch_handler/text_classifier.py) - This handler takes a text (string) as input and returns the classification text based on the model vocabulary
|
35 | 29 | * [Object Detector](https://github.com/pytorch/serve/blob/master/ts/torch_handler/object_detector.py) - This handler takes an image and returns list of detected classes and bounding boxes respectively
|
36 | 30 | * [Image Segmenter](https://github.com/pytorch/serve/blob/master/ts/torch_handler/image_segmenter.py)- This handler takes an image and returns output shape as [CL H W], CL - number of classes, H - height and W - width
|
| 31 | + |
| 32 | +## Examples |
| 33 | + |
| 34 | +* [HuggingFace Language Model](https://github.com/pytorch/serve/blob/master/examples/Huggingface_Transformers/Transformer_handler_generalized.py) - This handler takes an input sentence and can return sequence classifications, token classifications or Q&A answers |
| 35 | +* [Multi Modal Framework](https://github.com/pytorch/serve/blob/master/examples/MMF-activity-recognition/handler.py) - Build and deploy a classifier that combines text, audio and video input data |
| 36 | +* [Dual Translation Workflow](https://github.com/pytorch/serve/tree/master/examples/Workflows/nmt_tranformers_pipeline) - |
| 37 | +* [Model Zoo](model_zoo.md) - List of pre-trained model archives ready to be served for inference with TorchServe. |
| 38 | +* [Examples](https://github.com/pytorch/serve/tree/master/examples) - Many examples of how to package and deploy models with TorchServe |
| 39 | +* [Workflow Examples](https://github.com/pytorch/serve/tree/master/examples/Workflows) - Examples of how to compose models in a workflow with TorchServe |
| 40 | + |
| 41 | +## Advanced Features |
| 42 | + |
| 43 | +* [Advanced configuration](configuration.md) - Describes advanced TorchServe configurations. |
| 44 | +* [A/B test models](https://github.com/pytorch/serve/blob/master/docs/use_cases.md#serve-models-for-ab-testing) - A/B test your models for regressions before shipping them to production |
| 45 | +* [Custom Service](custom_service.md) - Describes how to develop custom inference services. |
| 46 | +* [Encrypted model serving](management_api.md/encrypted-model-serving) - S3 server side model encryption via KMS |
| 47 | +* [Snapshot serialization](https://github.com/pytorch/serve/blob/master/plugins/docs/ddb_endpoint.md) - Serialize model artifacts to AWS Dynamo DB |
| 48 | +* [Benchmarking and Profiling](https://github.com/pytorch/serve/tree/master/benchmarks#torchserve-model-server-benchmarking) - Use JMeter or Apache Bench to benchmark your models and TorchServe itself |
| 49 | +* [TorchServe on Kubernetes](https://github.com/pytorch/serve/blob/master/kubernetes/README.md#torchserve-on-kubernetes) - Demonstrates a Torchserve deployment in Kubernetes using Helm Chart supported in both Azure Kubernetes Service and Google Kubernetes service |
| 50 | +* [mlflow-torchserve](https://github.com/mlflow/mlflow-torchserve) - Deploy mlflow pipeline models into TorchServe |
| 51 | +* [Kubeflow pipelines](https://github.com/kubeflow/pipelines/tree/master/samples/contrib/pytorch-samples) - Kubeflow pipelines and Google Vertex AI Managed pipelines |
0 commit comments