From c9fadda54353f1b57c3dae9b7cbebda6f0767f8e Mon Sep 17 00:00:00 2001 From: Woosuk Kwon Date: Sun, 17 Dec 2023 02:28:02 -0800 Subject: [PATCH 1/2] [Minor] Fix xformers version (#2158) --- requirements.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/requirements.txt b/requirements.txt index e9023bde82a45..92ba0a716c45c 100644 --- a/requirements.txt +++ b/requirements.txt @@ -7,7 +7,7 @@ sentencepiece # Required for LLaMA tokenizer. numpy torch == 2.1.2 transformers >= 4.36.0 # Required for Mixtral. -xformers == 0.0.23 # Required for CUDA 12.1. +xformers == 0.0.23.post1 # Required for CUDA 12.1. fastapi uvicorn[standard] pydantic == 1.10.13 # Required for OpenAI server. From f8c688d746b32a54c249c4faf6fc73b83f53952a Mon Sep 17 00:00:00 2001 From: Woosuk Kwon Date: Sun, 17 Dec 2023 02:54:57 -0800 Subject: [PATCH 2/2] [Minor] Add Phi 2 to supported models (#2159) --- README.md | 2 +- docs/source/models/supported_models.rst | 4 ++-- tests/models/test_models.py | 4 ++-- 3 files changed, 5 insertions(+), 5 deletions(-) diff --git a/README.md b/README.md index d9cc6d26119c1..c6e6a3c7379db 100644 --- a/README.md +++ b/README.md @@ -65,7 +65,7 @@ vLLM seamlessly supports many Hugging Face models, including the following archi - Mixtral (`mistralai/Mixtral-8x7B-v0.1`, `mistralai/Mixtral-8x7B-Instruct-v0.1`, etc.) - MPT (`mosaicml/mpt-7b`, `mosaicml/mpt-30b`, etc.) - OPT (`facebook/opt-66b`, `facebook/opt-iml-max-30b`, etc.) -- Phi-1.5 (`microsoft/phi-1_5`, etc.) +- Phi (`microsoft/phi-1_5`, `microsoft/phi-2`, etc.) - Qwen (`Qwen/Qwen-7B`, `Qwen/Qwen-7B-Chat`, etc.) - Yi (`01-ai/Yi-6B`, `01-ai/Yi-34B`, etc.) diff --git a/docs/source/models/supported_models.rst b/docs/source/models/supported_models.rst index 44e4fe5ead988..c95b158e871fe 100644 --- a/docs/source/models/supported_models.rst +++ b/docs/source/models/supported_models.rst @@ -60,8 +60,8 @@ Alongside each architecture, we include some popular models that use it. - OPT, OPT-IML - :code:`facebook/opt-66b`, :code:`facebook/opt-iml-max-30b`, etc. * - :code:`PhiForCausalLM` - - Phi-1.5 - - :code:`microsoft/phi-1_5`, etc. + - Phi + - :code:`microsoft/phi-1_5`, :code:`microsoft/phi-2`, etc. * - :code:`QWenLMHeadModel` - Qwen - :code:`Qwen/Qwen-7B`, :code:`Qwen/Qwen-7B-Chat`, etc. diff --git a/tests/models/test_models.py b/tests/models/test_models.py index 95eabaafec811..e65c424c601a2 100644 --- a/tests/models/test_models.py +++ b/tests/models/test_models.py @@ -15,12 +15,12 @@ "EleutherAI/pythia-70m", "bigscience/bloom-560m", "mosaicml/mpt-7b", - "microsoft/phi-1_5", + "microsoft/phi-2", ] @pytest.mark.parametrize("model", MODELS) -@pytest.mark.parametrize("dtype", ["half"]) +@pytest.mark.parametrize("dtype", ["float"]) @pytest.mark.parametrize("max_tokens", [128]) def test_models( hf_runner,