Skip to content
Merged
Show file tree
Hide file tree
Changes from 6 commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion pylock.toml
Original file line number Diff line number Diff line change
Expand Up @@ -4227,4 +4227,4 @@ strategy = ["inherit_metadata", "static_urls"]
requires_python = "~=3.12"

[[tool.pdm.targets]]
requires_python = ">=3.10.0,<3.12"
requires_python = ">=3.10.0,<3.12"
25 changes: 15 additions & 10 deletions pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,13 @@ include = ["*"]
[tool.pdm]
distribution = true

[[tool.pdm.source]]
name = "torch"
type = "find_links"
#url = "https://download.pytorch.org/whl/cpu/torch_stable.html"
url = "https://download.pytorch.org/whl/cpu/torch/"
include_packages = ["torch"]


# ************************************************
# ********** Project Metadata **********
Expand Down Expand Up @@ -54,28 +61,26 @@ dependencies = [
"httpx[http2]<1.0.0",
"loguru",
"msgpack",
"numpy",
"numpy<2.0.0",
"pillow",
"protobuf",
"pydantic>=2.11.7",
"pydantic-settings>=2.0.0",
"pydub",
"pyyaml>=6.0.0",
"rich",
"sanic",
"transformers",
"uvloop>=0.18",
"librosa>=0.11.0",
"torch",
]

[project.optional-dependencies]
perf = [
"orjson",
"msgpack",
"msgspec",
"uvloop",
]
perf = ["orjson", "msgpack", "msgspec", "uvloop"]
recommended = [
"tiktoken>=0.11.0", # For OpenAI tokenizer
"blobfile>=3.1.0", # For OpenAI tokenizer
"tiktoken>=0.11.0", # For OpenAI tokenizer
"blobfile>=3.1.0", # For OpenAI tokenizer
]
dev = [
# build
Expand Down Expand Up @@ -118,7 +123,7 @@ dev = [
]

[dependency-groups]
dev = [ "guidellm[dev]" ]
dev = ["guidellm[dev]"]

[project.urls]
homepage = "https://github.com/vllm-project/guidellm"
Expand Down
3 changes: 3 additions & 0 deletions src/guidellm/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,8 @@
import logging
import os

from datasets import config

with (
open(os.devnull, "w") as devnull, # noqa: PTH123
contextlib.redirect_stderr(devnull),
Expand All @@ -19,6 +21,7 @@
os.environ["TOKENIZERS_PARALLELISM"] = "false" # Silence warnings for tokenizers
hf_logging.set_verbosity_error()
logging.getLogger("transformers").setLevel(logging.ERROR)
config.USE_AUDIO_DECODE = False

from .logger import configure_logger, logger
from .settings import (
Expand Down
Loading
Loading