Skip to content

docs: Add multimodal architecture doc #1096

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Binary file added docs/images/multimodal_epd_disagg_workflow.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
35 changes: 35 additions & 0 deletions docs/multimodal/architecture.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,35 @@

# Dynamo Multimodal Architecture

## Motivation

Multimodal models enhance the capabilities of large language models (LLMs) by enabling them to process diverse input types such as text, images, and audio. While powerful, these models pose significant challenges for efficient and scalable inference.

To address this, the paper [Efficiently Serving Large Multimodal Models Using EPD Disaggregation](https://arxiv.org/pdf/2501.05460) introduces key optimizations—most notably, the **EPD Disaggregation** technique. This approach decomposes the inference process into three distinct stages: **Encode**, **Prefill**, and **Decode**. By separating these phases, the system can unlock higher throughput, better resource utilization, and improved scalability.

The Dynamo multimodal architecture adopts and extends this EPD disaggregation strategy to support efficient serving of large multimodal models.

## High level architecture

Dynamo has already demonstrated [prefill-decode (PD) disaggregation](../disagg_serving.md) for efficient LLM serving. Building on that foundation, this proposal introduces a dedicated **encode worker** to handle the vision encoding step, fully separating the Encode, Prefill, and Decode phases.
The diagram below illustrates the proposed workflow, based on the LLAVA model ( image + text → text) with vLLM.

![](../images/multimodal_epd_disagg_workflow.png "Multimodal EPD Disaggregation Workflow")
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue

Add alt text for accessibility. The image tag ![](../images/multimodal_epd_disagg_workflow.png "Multimodal EPD Disaggregation Workflow") lacks alt text. Replace with:

![Multimodal EPD Disaggregation Workflow](../images/multimodal_epd_disagg_workflow.png "Multimodal EPD Disaggregation Workflow")
🧰 Tools
🪛 markdownlint-cli2 (0.17.2)

17-17: Images should have alternate text (alt text)
null

(MD045, no-alt-text)

🤖 Prompt for AI Agents
In docs/multimodal/architecture.md at line 17, the image markdown lacks alt text
which is important for accessibility. Add descriptive alt text by changing the
image tag from ![](../images/multimodal_epd_disagg_workflow.png "Multimodal EPD
Disaggregation Workflow") to include alt text like ![Multimodal EPD
Disaggregation Workflow](../images/multimodal_epd_disagg_workflow.png
"Multimodal EPD Disaggregation Workflow").


### Components

- Client / Frontend: Sends a request containing a text prompt and an image URL or path to the serving endpoint.

- Processor: Handles pre-processing of the input prompt and post-processing of the LLM output.

- Decode Worker (vLLM Worker):
- In aggregation mode, this worker handles all stages of inference.
- In disaggregated mode, it focuses on decoding and issues a remote prefill request to the Prefill Queue, where the Prefill Worker picks it up.

- Prefill Worker:
- Handles the prefill stage of inference.
- In multimodal workflows, it forwards the image URL to the Encode Worker, waits for the returned embeddings, and then proceeds with the prefill using both the text and image features.

- Encode Worker:
- Encodes multimodal inputs (e.g., images) into embeddings.
- In this example, it converts an image into vision embeddings and sends them to the Prefill Worker for the prefill stage.
Loading