Skip to content

Commit 03e9ec4

Browse files
committed
[Misc][ModelScope] Change to use runtime VLLM_USE_MODELSCOPE
Signed-off-by: Mengqing Cao <cmq0113@163.com>
1 parent a859320 commit 03e9ec4

File tree

4 files changed

+10
-10
lines changed

4 files changed

+10
-10
lines changed

vllm/model_executor/model_loader/default_loader.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,8 +11,8 @@
1111
from torch import nn
1212
from transformers.utils import SAFE_WEIGHTS_INDEX_NAME
1313

14+
from vllm import envs
1415
from vllm.config import LoadConfig, LoadFormat, ModelConfig, VllmConfig
15-
from vllm.envs import VLLM_USE_MODELSCOPE
1616
from vllm.logger import init_logger
1717
from vllm.model_executor.model_loader.base_loader import BaseModelLoader
1818
from vllm.model_executor.model_loader.utils import (
@@ -64,7 +64,7 @@ def _maybe_download_from_modelscope(
6464
6565
Returns the path to the downloaded model, or None if the model is not
6666
downloaded from ModelScope."""
67-
if VLLM_USE_MODELSCOPE:
67+
if envs.VLLM_USE_MODELSCOPE:
6868
# download model from ModelScope hub,
6969
# lazy import so that modelscope is not required for normal use.
7070
# pylint: disable=C.

vllm/transformers_utils/__init__.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
11
# SPDX-License-Identifier: Apache-2.0
22

3-
from vllm.envs import VLLM_USE_MODELSCOPE
3+
from vllm import envs
44

5-
if VLLM_USE_MODELSCOPE:
5+
if envs.VLLM_USE_MODELSCOPE:
66
try:
77
# Patch here, before each import happens
88
import modelscope

vllm/transformers_utils/config.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@
2424
MODEL_FOR_CAUSAL_LM_MAPPING_NAMES)
2525
from transformers.utils import CONFIG_NAME as HF_CONFIG_NAME
2626

27-
from vllm.envs import VLLM_USE_MODELSCOPE
27+
from vllm import envs
2828
from vllm.logger import init_logger
2929
# yapf conflicts with isort for this block
3030
# yapf: disable
@@ -45,7 +45,7 @@
4545
from vllm.transformers_utils.utils import check_gguf_file
4646
from vllm.utils import resolve_obj_by_qualname
4747

48-
if VLLM_USE_MODELSCOPE:
48+
if envs.VLLM_USE_MODELSCOPE:
4949
from modelscope import AutoConfig
5050
else:
5151
from transformers import AutoConfig
@@ -130,7 +130,7 @@ def lookup_files() -> list[str]:
130130
]
131131
# if model is remote, use hf_hub api to list files
132132
try:
133-
if VLLM_USE_MODELSCOPE:
133+
if envs.VLLM_USE_MODELSCOPE:
134134
from vllm.transformers_utils.utils import (
135135
modelscope_list_repo_files)
136136
return modelscope_list_repo_files(repo_id,
@@ -768,7 +768,7 @@ def get_hf_image_processor_config(
768768
**kwargs,
769769
) -> dict[str, Any]:
770770
# ModelScope does not provide an interface for image_processor
771-
if VLLM_USE_MODELSCOPE:
771+
if envs.VLLM_USE_MODELSCOPE:
772772
return dict()
773773
# Separate model folder from file path for GGUF models
774774
if check_gguf_file(model):

vllm/transformers_utils/tokenizer.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
from transformers import (AutoTokenizer, PreTrainedTokenizer,
1414
PreTrainedTokenizerFast)
1515

16-
from vllm.envs import VLLM_USE_MODELSCOPE
16+
from vllm import envs
1717
from vllm.logger import init_logger
1818
from vllm.lora.request import LoRARequest
1919
from vllm.transformers_utils.tokenizer_base import (TokenizerBase,
@@ -168,7 +168,7 @@ def get_tokenizer(
168168
) -> AnyTokenizer:
169169
"""Gets a tokenizer for the given model name via HuggingFace or ModelScope.
170170
"""
171-
if VLLM_USE_MODELSCOPE:
171+
if envs.VLLM_USE_MODELSCOPE:
172172
# download model from ModelScope hub,
173173
# lazy import so that modelscope is not required for normal use.
174174
# pylint: disable=C.

0 commit comments

Comments
 (0)