Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update index.md #3

Merged
merged 1 commit into from
Jul 21, 2022
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Update index.md
edits
  • Loading branch information
samyam authored Jul 21, 2022
commit 059506a63600cae3522872b884d55b8e6a5aec53
14 changes: 7 additions & 7 deletions docs/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,13 +24,13 @@ title: "Latest News"

# DeepSpeed: Extreme Speed and Scale for DL Training and Inference

DeepSpeed is an easy-to-use deep learning optimization suite that enables unprecedented scale and speed for Deep Learning Training and Inference.
DeepSpeed is an easy-to-use deep learning optimization suite that enables unprecedented scale and speed for Deep Learning Training and Inference. With DeepSpeed, you can:

- DeepSpeed empowers data scientists to Train/Inference dense or sparse models with billions or trillions of parameters
- Achieve excellent system throughput and efficiently scale to thousands of GPUs
- Train/Inference on resource constrained GPU systems
- Achieve unprecedented low latency and high thoughput for inference
- Achieve extreme compression for an unparalleled inference latency and model size reduction with low costs
<p style="text-align: center;">Train/Inference dense or sparse models with billions or trillions of parameters</p>
<p style="text-align: center;">Achieve excellent system throughput and efficiently scale to thousands of GPUs</p>
<p style="text-align: center;">Train/Inference on resource constrained GPU systems</p>
<p style="text-align: center;">Achieve unprecedented low latency and high thoughput for inference</p>
<p style="text-align: center;">Achieve extreme compression for an unparalleled inference latency and model size reduction with low costs</p>


## Three main innovation pillars
Expand All @@ -50,7 +50,7 @@ DeepSpeed brings together innovations in parallelism technology such as tensor,

### DeepSpeed-Compression:

To further increase the infrence efficency, DeepSpeed provides a new feature, that offers an easy-to-use and flexible-to-compose compression library for researchers and practitioners to compress their models while delivering faster speed, smaller model size, and significantly reduced compression cost. Meanwhile, new innovations, like ZeroQuant and XTC, are included under the DeepSpeed-Compression pillar. (See [here](https://www.deepspeed.ai/tutorials/model-compression/) for more details)
To further increase the inference efficency, DeepSpeed offers easy-to-use and flexible-to-compose compression techniques for researchers and practitioners to compress their models while delivering faster speed, smaller model size, and significantly reduced compression cost. Moreover, SoTA innovations on compression like ZeroQuant and XTC are included under the DeepSpeed-Compression pillar. (See [here](https://www.deepspeed.ai/tutorials/model-compression/) for more details)

## DeepSpeed Software Suite

Expand Down