Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[TPU][Quantization] TPU
W8A8
#11785[TPU][Quantization] TPU
W8A8
#11785Changes from 1 commit
3b0c8a6
36fc1db
d83c04c
af9d0f4
0f9fd21
7b3203f
bf50fa4
226ef52
bb7c741
cf842bd
67039bc
0695f77
11cf82f
e016e38
717b859
c848735
1539915
f00412a
b0a6b70
e812d7e
764dda1
87b2ae6
e813ff8
8cfaa1b
bbc9741
eb3f39e
bb2fbe1
14ccb90
4092be2
1aaa628
48aa54b
4efe915
e98b79c
5a89668
57cbf5c
3451c4d
0c2e62a
172c9ca
938ca81
9e18911
5f58ec7
af9f298
6fe2f62
f2c0beb
8b29718
1e2a373
2a359ef
f7e8975
0d4c3fd
57340d2
38291d5
ead1e94
cea5e54
940ddde
84a5b29
a1d7b4a
2b4ecfd
186c108
7e8598a
de773cd
3a53d7d
0be5f69
cb69ba7
3896f6c
33e1e13
dde72d6
d7a9c93
db9f795
09ad869
b74c88a
da4369e
5ddcac2
f353c43
File filter
Filter by extension
Conversations
Jump to
There are no files selected for viewing
Check failure on line 4 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 4 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 4 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 4 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 4 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 5 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 5 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 5 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 5 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 5 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (F401)
Check failure on line 40 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (E501)
Check failure on line 40 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (E501)
Check failure on line 40 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (E501)
Check failure on line 40 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (E501)
Check failure on line 40 in vllm/model_executor/layers/quantization/kernels/scaled_mm/xla.py
Ruff (E501)