-
Notifications
You must be signed in to change notification settings - Fork 183
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Transformers save/load compatibility and inference kernels #3
Changes from 1 commit
e159c7f
9ec6b70
395a9f6
8e80100
4cabbc3
2c97a9f
7864596
3f66363
07986d9
56359db
6ad6a6b
46be31d
6ee4353
9e917f1
e281962
8c35e89
a795ee6
10c0a7e
c41f09e
ed1430f
eb89781
96301f7
c19a5f6
0013d8e
89d6085
444f788
0f5483a
a9e65bb
27e3855
0f37ea6
e3b623f
fe9a9f2
164c5d1
db0d5a2
0d7e1af
9f83593
990d5b2
b0b59f5
fd8395f
e519193
9e89f61
e673e0b
0d69cf6
59e1f3c
e98ed7c
c245c5e
147ed79
a5ae331
11e5dfd
e5e9ee9
48d6ddf
51d664b
8979ae1
4f31eae
c9cf936
d0f6ed4
6cc2756
7476833
eb8c2cd
1db5115
7f7e853
5b3a5d2
1804499
07f72b6
bf0880f
d7c4561
823db17
22a7994
b906bfd
6a6ebd3
3937640
c643fec
d67d119
c31d532
2d0cae8
3deeab2
aca05dd
cfa5e4a
9498bf3
7c6d234
426a7b6
78cc9a8
1278164
2dbd188
935347e
7b8faf8
33b0464
ead1c00
88d9a93
28d70f8
b31a3fc
d9f6b25
503ff40
26ff8b0
09a7810
c434d42
788c289
9fdf0a6
f2ef38b
7342655
989d5d8
5d4f4f3
2a32c0a
f019b4e
d90c43b
e06a789
098363a
d7b6dfa
4bd67b9
d44c29d
79706d0
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
- Loading branch information
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,6 +1,6 @@ | ||
[metadata] | ||
name = aqlm | ||
version = 1.0.0dev6 | ||
version = 1.0.0dev7 | ||
author = Vage Egiazarian, Andrei Panferov and Elias Frantar | ||
author_email = vahe527887@yandex.ru | ||
description = Efficiently run models quantized with AQLM | ||
|
@@ -33,8 +33,12 @@ python_requires = >=3.10 | |
install_requires = | ||
torch>=2.1 | ||
transformers>=4.37 | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Nit/recommendation: upper bound transformer version, e.g. >=0.37,<0.38 Why: your benchmarking code relies on HF internals that often break with major patches There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Can't make it work. The version range isn't properly parsed for some reason |
||
triton>=2.1 | ||
[options.extras_require] | ||
gpu = | ||
triton>=2.1 | ||
cpu = | ||
numba>=0.56.4 | ||
scipy>=1.11.3 | ||
dev = | ||
pytest==6.2.5 | ||
pytest-forked | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It is your decision. but I would recommend "AQLM paper authors" if you have no preference
If you chose this deliberately, I have no objections.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
"AQLM paper authors" makes perfect sense. Will do that
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Done