Skip to content

Commit

Permalink
Intel ci (#2630)
Browse files Browse the repository at this point in the history
* Intel CI ?

* Let's try non sharded gemma.

* Snapshot rename

* Apparently container can be gone already.
  • Loading branch information
Narsil authored Oct 10, 2024
1 parent d912f0b commit 3dbdf63
Show file tree
Hide file tree
Showing 6 changed files with 9 additions and 6 deletions.
6 changes: 3 additions & 3 deletions .github/workflows/build.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -75,10 +75,10 @@ jobs:
export label_extension="-intel-cpu"
export docker_devices="none"
export docker_volume="/mnt/cache"
export runs_on="ubuntu-latest"
# export runs_on="aws-highmemory-32-plus-priv"
# export runs_on="ubuntu-latest"
export runs_on="aws-highmemory-32-plus-priv"
export platform="cpu"
export extra_pytest="-k test_flash_llama_load"
export extra_pytest="-k test_flash_gemma_simple"
;;
esac
echo $dockerfile
Expand Down
5 changes: 4 additions & 1 deletion integration-tests/conftest.py
Original file line number Diff line number Diff line change
Expand Up @@ -572,7 +572,10 @@ def docker_launcher(
print(container_output, file=sys.stderr)

finally:
container.remove()
try:
container.remove()
except Exception:
pass

if DOCKER_IMAGE is not None:
return docker_launcher
Expand Down
2 changes: 1 addition & 1 deletion integration-tests/models/test_flash_gemma.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ async def flash_gemma(flash_gemma_handle):
@pytest.mark.release
@pytest.mark.asyncio
@pytest.mark.private
async def test_flash_gemma(flash_gemma, response_snapshot):
async def test_flash_gemma_simple(flash_gemma, response_snapshot):
response = await flash_gemma.generate(
"Test request", max_new_tokens=10, decoder_input_details=True
)
Expand Down
2 changes: 1 addition & 1 deletion integration-tests/models/test_flash_llama.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ async def flash_llama(flash_llama_handle):

@pytest.mark.asyncio
@pytest.mark.private
async def test_flash_llama(flash_llama, response_snapshot):
async def test_flash_llama_simple(flash_llama, response_snapshot):
response = await flash_llama.generate(
"Test request", max_new_tokens=10, decoder_input_details=True
)
Expand Down

0 comments on commit 3dbdf63

Please sign in to comment.