Skip to content

Commit 4e142ba

Browse files
authored
[GuideLLM Refactor] Data pipelines rework and multimodal support (#384)
…icated combinations ## Summary <!-- Include a short paragraph of the changes introduced in this PR. If this PR requires additional context or rationale, explain why the changes are necessary. --> ## Details <!-- Provide a detailed list of all changes introduced in this pull request. --> - [ ] ## Test Plan <!-- List the steps needed to test this PR. --> - ## Related Issues <!-- Link any relevant issues that this PR addresses. --> - Resolves # --- - [ ] "I certify that all code in this PR is my own, except as noted below." ## Use of AI - [ ] Includes AI-assisted code completion - [ ] Includes code generated by an AI application - [ ] Includes AI-generated tests (NOTE: AI written tests should have a docstring that includes `## WRITTEN BY AI ##`)
2 parents b162fb3 + 16f981c commit 4e142ba

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

83 files changed

+6792
-5900
lines changed

pylock.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4227,4 +4227,4 @@ strategy = ["inherit_metadata", "static_urls"]
42274227
requires_python = "~=3.12"
42284228

42294229
[[tool.pdm.targets]]
4230-
requires_python = ">=3.10.0,<3.12"
4230+
requires_python = ">=3.10.0,<3.12"

pyproject.toml

Lines changed: 15 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,13 @@ include = ["*"]
1313
[tool.pdm]
1414
distribution = true
1515

16+
[[tool.pdm.source]]
17+
name = "torch"
18+
type = "find_links"
19+
#url = "https://download.pytorch.org/whl/cpu/torch_stable.html"
20+
url = "https://download.pytorch.org/whl/cpu/torch/"
21+
include_packages = ["torch"]
22+
1623

1724
# ************************************************
1825
# ********** Project Metadata **********
@@ -54,29 +61,25 @@ dependencies = [
5461
"httpx[http2]<1.0.0",
5562
"loguru",
5663
"msgpack",
57-
"numpy",
64+
"numpy<2.0.0",
5865
"pillow",
5966
"protobuf",
6067
"pydantic>=2.11.7",
6168
"pydantic-settings>=2.0.0",
69+
"pydub",
6270
"pyyaml>=6.0.0",
6371
"rich",
6472
"sanic",
6573
"transformers",
6674
"uvloop>=0.18",
75+
"librosa>=0.11.0",
76+
"torch",
6777
]
6878

6979
[project.optional-dependencies]
70-
perf = [
71-
"orjson",
72-
"msgpack",
73-
"msgspec",
74-
"uvloop",
75-
]
76-
recommended = [
77-
"tiktoken>=0.11.0", # For OpenAI tokenizer
78-
"blobfile>=3.1.0", # For OpenAI tokenizer
79-
]
80+
perf = ["orjson", "msgpack", "msgspec", "uvloop"]
81+
openai = ["tiktoken>=0.11.0", "blobfile>=3.1.0"]
82+
recommended = ["guidellm[perf,openai]"]
8083
dev = [
8184
# build
8285
"build>=1.0.0",
@@ -118,7 +121,7 @@ dev = [
118121
]
119122

120123
[dependency-groups]
121-
dev = [ "guidellm[dev]" ]
124+
dev = ["guidellm[dev]"]
122125

123126
[project.urls]
124127
homepage = "https://github.com/vllm-project/guidellm"

src/guidellm/__init__.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,8 @@
77
import logging
88
import os
99

10+
from datasets import config
11+
1012
with (
1113
open(os.devnull, "w") as devnull, # noqa: PTH123
1214
contextlib.redirect_stderr(devnull),
@@ -19,6 +21,7 @@
1921
os.environ["TOKENIZERS_PARALLELISM"] = "false" # Silence warnings for tokenizers
2022
hf_logging.set_verbosity_error()
2123
logging.getLogger("transformers").setLevel(logging.ERROR)
24+
config.USE_AUDIO_DECODE = False
2225

2326
from .logger import configure_logger, logger
2427
from .settings import (

0 commit comments

Comments
 (0)