Skip to content

Commit ff38f0a

Browse files
varun-sundar-rabindranathVarun Sundar Rabindranath
and
Varun Sundar Rabindranath
authored
[CI/Build] LoRA: Delete long context tests (#15503)
Signed-off-by: Varun Sundar Rabindranath <varun@neuralmagic.com> Co-authored-by: Varun Sundar Rabindranath <varun@neuralmagic.com>
1 parent a5cfbab commit ff38f0a

File tree

2 files changed

+1
-304
lines changed

2 files changed

+1
-304
lines changed

.buildkite/test-pipeline.yaml

+1-3
Original file line numberDiff line numberDiff line change
@@ -287,7 +287,7 @@ steps:
287287
source_file_dependencies:
288288
- vllm/lora
289289
- tests/lora
290-
command: pytest -v -s lora --shard-id=$$BUILDKITE_PARALLEL_JOB --num-shards=$$BUILDKITE_PARALLEL_JOB_COUNT --ignore=lora/test_long_context.py --ignore=lora/test_chatglm3_tp.py --ignore=lora/test_llama_tp.py --ignore=lora/test_minicpmv_tp.py --ignore=lora/test_transfomers_model.py
290+
command: pytest -v -s lora --shard-id=$$BUILDKITE_PARALLEL_JOB --num-shards=$$BUILDKITE_PARALLEL_JOB_COUNT --ignore=lora/test_chatglm3_tp.py --ignore=lora/test_llama_tp.py --ignore=lora/test_minicpmv_tp.py --ignore=lora/test_transfomers_model.py
291291
parallelism: 4
292292

293293
- label: PyTorch Fullgraph Smoke Test # 9min
@@ -592,8 +592,6 @@ steps:
592592
# FIXIT: find out which code initialize cuda before running the test
593593
# before the fix, we need to use spawn to test it
594594
- export VLLM_WORKER_MULTIPROC_METHOD=spawn
595-
# This test runs llama 13B, so it is required to run on 4 GPUs.
596-
- pytest -v -s -x lora/test_long_context.py
597595
# There is some Tensor Parallelism related processing logic in LoRA that
598596
# requires multi-GPU testing for validation.
599597
- pytest -v -s -x lora/test_chatglm3_tp.py

tests/lora/test_long_context.py

-301
This file was deleted.

0 commit comments

Comments
 (0)