Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sparse attn triton v1.0 support + torch1.8 test runner #1374

Merged
merged 23 commits into from
Sep 21, 2021
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
23 commits
Select commit Hold shift + click to select a range
18a8dff
let the sparse tests run
Jun 10, 2021
80dbe2f
fixing the sparse-APIs to use the latest triton version
Jun 29, 2021
be55c85
update with assert and some fixes
jeffra Sep 15, 2021
9e2dee4
add torch18 tests and fix sparse-attn checks
jeffra Sep 15, 2021
e702866
turn back on tests
jeffra Sep 16, 2021
54752d0
use relative paths for megatron jsons
jeffra Sep 14, 2021
741f8b5
factor out relative path for unit test files
jeffra Sep 14, 2021
626a51e
set test path
jeffra Sep 16, 2021
318ae95
refactor sparse attn imports
jeffra Sep 16, 2021
9963435
Merge branch 'master' into reyazda/test-sparse-v2
jeffra Sep 16, 2021
49514bd
fix relative import
jeffra Sep 16, 2021
61bf986
rename test_path so pytest doesn't think its a test
jeffra Sep 16, 2021
35245fe
skip test_configurable_parallel for now until fixed
jeffra Sep 16, 2021
0d28040
moe fix
jeffra Sep 17, 2021
d1615ff
fixes random connection reset test failures for some unit tests
jeffra Sep 17, 2021
940abc6
resolve the TK with correct setting based on dtype and block size
Sep 17, 2021
bd13efd
Merge branch 'reyazda/test-sparse-v2' of github.com:microsoft/DeepSpe…
Sep 17, 2021
7d9c5fb
fix megatron regression, add moe unit test, fix moe ckpt comparison
jeffra Sep 17, 2021
7733f00
add sparse-attn skip if not compatible
jeffra Sep 17, 2021
8001124
Merge branch 'master' into reyazda/test-sparse-v2
jeffra Sep 18, 2021
08695e2
skip moe ckpt test if old torch
jeffra Sep 20, 2021
1324589
turn back on test_configurable_parallel
jeffra Sep 20, 2021
3beb9fd
tear down torch dist pg when test completes
jeffra Sep 20, 2021
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
29 changes: 28 additions & 1 deletion .github/workflows/main.yml
Original file line number Diff line number Diff line change
Expand Up @@ -37,4 +37,31 @@ jobs:
- name: Unit tests
run: |
if [[ -d ./torch-extensions ]]; then rm -rf ./torch-extensions; fi
TORCH_EXTENSIONS_DIR=./torch-extensions pytest --durations=0 --forked --verbose tests/unit/
cd tests
TORCH_EXTENSIONS_DIR=./torch-extensions pytest --durations=0 --forked --verbose unit/

nv-torch18-v100:
runs-on: [self-hosted, nvidia, torch18, v100]

steps:
- uses: actions/checkout@v2

- name: environment
run: |
nvidia-smi
which python
python --version
which nvcc
nvcc --version
python -c "import torch; print('torch:', torch.__version__, torch)"
python -c "import torch; print('CUDA available:', torch.cuda.is_available())"
- name: Install deepspeed
run: |
pip install .[dev,1bit]
ds_report
- name: Unit tests
run: |
unset TORCH_CUDA_ARCH_LIST # only jit compile for current arch
if [[ -d ./torch-extensions ]]; then rm -rf ./torch-extensions; fi
cd tests
TORCH_EXTENSIONS_DIR=./torch-extensions pytest --durations=0 --forked --verbose unit/
2 changes: 0 additions & 2 deletions deepspeed/ops/sparse_attention/__init__.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,4 @@
from .sparsity_config import SparsityConfig, DenseSparsityConfig, FixedSparsityConfig, VariableSparsityConfig, BigBirdSparsityConfig, BSLongformerSparsityConfig
from .softmax import Softmax
from .matmul import MatMul
from .sparse_self_attention import SparseSelfAttention
from .bert_sparse_self_attention import BertSparseSelfAttention
from .sparse_attention_utils import SparseAttentionUtils
Loading