Skip to content

Commit

Permalink
formats
Browse files Browse the repository at this point in the history
Signed-off-by: Keyun Tong <tongkeyun@gmail.com>
  • Loading branch information
youngkent committed Jan 29, 2025
1 parent 54147cb commit ec218b0
Show file tree
Hide file tree
Showing 3 changed files with 10 additions and 9 deletions.
12 changes: 6 additions & 6 deletions vllm/transformers_utils/tokenizer.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,11 +12,11 @@
from vllm.envs import VLLM_USE_MODELSCOPE
from vllm.logger import init_logger
from vllm.lora.request import LoRARequest
from vllm.transformers_utils.tokenizer_base import (TokenizerBase,
TokenizerRegistry)
from vllm.transformers_utils.tokenizers import MistralTokenizer
from vllm.transformers_utils.utils import check_gguf_file
from vllm.utils import make_async
from vllm.transformers_utils.tokenizer_base import (TokenizerBase,
TokenizerRegistry)

logger = init_logger(__name__)

Expand Down Expand Up @@ -188,10 +188,10 @@ def get_tokenizer(
revision=revision)
elif tokenizer_mode == "custom":
tokenizer = TokenizerRegistry.get_tokenizer(str(tokenizer_name),
*args,
revision=revision,
download_dir=download_dir,
**kwargs)
*args,
revision=revision,
download_dir=download_dir,
**kwargs)
else:
try:
tokenizer = AutoTokenizer.from_pretrained(
Expand Down
5 changes: 3 additions & 2 deletions vllm/transformers_utils/tokenizer_base.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
from abc import ABC, abstractmethod
from typing import List, Union, Dict, Optional, TYPE_CHECKING, Any, Tuple
import importlib
from abc import ABC, abstractmethod
from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union

if TYPE_CHECKING:
from vllm.entrypoints.chat_utils import ChatCompletionMessageParam

Expand Down
2 changes: 1 addition & 1 deletion vllm/transformers_utils/tokenizers/mistral.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,9 +16,9 @@
SentencePieceTokenizer)
from mistral_common.tokens.tokenizers.tekken import (SpecialTokenPolicy,
Tekkenizer)
from vllm.transformers_utils.tokenizer_base import TokenizerBase

from vllm.logger import init_logger
from vllm.transformers_utils.tokenizer_base import TokenizerBase
from vllm.utils import is_list_of

if TYPE_CHECKING:
Expand Down

0 comments on commit ec218b0

Please sign in to comment.