Commit 81da403
Enhance SmoothQuant tuning structure. (#1109)
* enhance sq tuning
Signed-off-by: Xin He <xin3.he@intel.com>
* Support the tuning of smooth quant' alpha in strategy layer (#1112)
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* added more UTs
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* fixed ut
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* fixed ut
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* enable sq tuning for both quant_level is auto or 1
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* fix accuracy issue
Signed-off-by: Xin He <xin3.he@intel.com>
* fix UT
Signed-off-by: Xin He <xin3.he@intel.com>
* fix alpha=auto
Signed-off-by: Xin He <xin3.he@intel.com>
* support sq tuning for both auto and O1
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* fixed the typo
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* rename func name in ut
Signed-off-by: Xin He <xin3.he@intel.com>
* remove duplicate Linear if Linear is wrapped by Linear
Signed-off-by: Xin He <xin3.he@intel.com>
* refactor tensorflow interface
* adjust the pre-optimization and sq order for ort
* updated ort ut
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* fix pylint and docstyle
Signed-off-by: Xin He <xin3.he@intel.com>
* add sketch for ort tune sq alpha
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* correct the calib_iter
Signed-off-by: yiliu30 <yi4.liu@intel.com>
* fix tensorflow UT and int8 acc issue
* fix ut
Signed-off-by: Xin He <xin3.he@intel.com>
---------
Signed-off-by: Xin He <xin3.he@intel.com>
Signed-off-by: yiliu30 <yi4.liu@intel.com>
Co-authored-by: Yi30 <106061964+yiliu30@users.noreply.github.com>
Co-authored-by: yiliu30 <yi4.liu@intel.com>
Co-authored-by: spycsh <sihan.chen@intel.com>1 parent 9ff7f01 commit 81da403
File tree
19 files changed
+861
-577
lines changed- neural_compressor
- adaptor
- tf_utils
- torch_utils
- algorithm
- strategy
- utils
- test
- adaptor/onnxrt_adaptor
- algorithm
- quantization
19 files changed
+861
-577
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
37 | 37 | | |
38 | 38 | | |
39 | 39 | | |
| 40 | + | |
40 | 41 | | |
41 | 42 | | |
42 | 43 | | |
| |||
157 | 158 | | |
158 | 159 | | |
159 | 160 | | |
160 | | - | |
161 | | - | |
| 161 | + | |
| 162 | + | |
| 163 | + | |
| 164 | + | |
| 165 | + | |
| 166 | + | |
| 167 | + | |
162 | 168 | | |
163 | 169 | | |
164 | 170 | | |
165 | 171 | | |
166 | 172 | | |
167 | 173 | | |
168 | | - | |
169 | 174 | | |
170 | 175 | | |
171 | 176 | | |
172 | 177 | | |
173 | 178 | | |
174 | 179 | | |
| 180 | + | |
175 | 181 | | |
176 | 182 | | |
177 | 183 | | |
| |||
180 | 186 | | |
181 | 187 | | |
182 | 188 | | |
183 | | - | |
184 | | - | |
185 | | - | |
| 189 | + | |
| 190 | + | |
| 191 | + | |
| 192 | + | |
| 193 | + | |
| 194 | + | |
| 195 | + | |
186 | 196 | | |
| 197 | + | |
| 198 | + | |
| 199 | + | |
187 | 200 | | |
| 201 | + | |
| 202 | + | |
| 203 | + | |
| 204 | + | |
| 205 | + | |
| 206 | + | |
| 207 | + | |
| 208 | + | |
| 209 | + | |
| 210 | + | |
| 211 | + | |
188 | 212 | | |
189 | 213 | | |
190 | 214 | | |
| |||
201 | 225 | | |
202 | 226 | | |
203 | 227 | | |
| 228 | + | |
| 229 | + | |
| 230 | + | |
| 231 | + | |
| 232 | + | |
| 233 | + | |
| 234 | + | |
| 235 | + | |
204 | 236 | | |
205 | 237 | | |
206 | 238 | | |
| |||
630 | 662 | | |
631 | 663 | | |
632 | 664 | | |
| 665 | + | |
| 666 | + | |
| 667 | + | |
| 668 | + | |
| 669 | + | |
633 | 670 | | |
634 | 671 | | |
635 | 672 | | |
| |||
0 commit comments