1717
1818MODELS = [
1919 "google/gemma-2-2b-it" ,
20- "meta-llama/Llama-3.2-1B" ,
20+ "meta-llama/Llama-3.2-1B-Instruct " ,
2121]
2222
2323TARGET_TEST_SUITE = os .environ .get ("TARGET_TEST_SUITE" , "L4" )
@@ -96,12 +96,12 @@ def test_models(
9696 "test_suite" , [
9797 ("facebook/opt-125m" , "ray" , "" , "L4" ),
9898 ("facebook/opt-125m" , "mp" , "" , "L4" ),
99- ("meta-llama/Llama-2-7b-hf " , "ray" , "" , "L4" ),
100- ("meta-llama/Llama-2-7b-hf " , "mp" , "" , "L4" ),
99+ ("meta-llama/Llama-3.2-1B-Instruct " , "ray" , "" , "L4" ),
100+ ("meta-llama/Llama-3.2-1B-Instruct " , "mp" , "" , "L4" ),
101101 ("facebook/opt-125m" , "ray" , "" , "A100" ),
102102 ("facebook/opt-125m" , "mp" , "" , "A100" ),
103103 ("facebook/opt-125m" , "mp" , "FLASHINFER" , "A100" ),
104- ("meta-llama/Meta- Llama-3-8B " , "ray" , "FLASHINFER" , "A100" ),
104+ ("meta-llama/Llama-3.2-1B-Instruct " , "ray" , "FLASHINFER" , "A100" ),
105105 ])
106106def test_models_distributed (
107107 hf_runner ,
@@ -116,7 +116,7 @@ def test_models_distributed(
116116 if test_suite != TARGET_TEST_SUITE :
117117 pytest .skip (f"Skip test for { test_suite } " )
118118
119- if model == "meta-llama/Llama-2-7b-hf " and distributed_executor_backend == "ray" and attention_backend == "" and test_suite == "L4" : # noqa
119+ if model == "meta-llama/Llama-3.2-1B-Instruct " and distributed_executor_backend == "ray" and attention_backend == "" and test_suite == "L4" : # noqa
120120 # test ray adag
121121 os .environ ['VLLM_USE_RAY_SPMD_WORKER' ] = "1"
122122 os .environ ['VLLM_USE_RAY_COMPILED_DAG' ] = "1"
0 commit comments