Skip to content

Conversation

@n1ck-guo
Copy link
Contributor

Type of Change

documentation

Description

add layer wise quantization doc in quantization_weight_only

Expected Behavior & Potential Risk

None

How has this PR been tested?

how to reproduce the test (including hardware information)

Dependency Change?

None

Signed-off-by: Guo, Heng <heng.guo@intel.com>
@n1ck-guo n1ck-guo requested a review from chensuyue October 11, 2023 01:20
@chensuyue
Copy link
Contributor

Please add support matrix for lwq.

Signed-off-by: Guo, Heng <heng.guo@intel.com>
@n1ck-guo n1ck-guo requested a review from yintong-lu October 16, 2023 05:16
@n1ck-guo n1ck-guo merged commit 1163603 into master Oct 16, 2023
@n1ck-guo n1ck-guo deleted the hengguo/lwq_doc branch October 16, 2023 05:22
bmyrcha pushed a commit that referenced this pull request Oct 24, 2023
* add lwq doc in quantization_weight_only

Signed-off-by: Guo, Heng <heng.guo@intel.com>
Signed-off-by: bmyrcha <bartosz.myrcha@intel.com>
chensuyue pushed a commit to chensuyue/lpot that referenced this pull request Feb 21, 2024
…quantization (intel#1311)

* Add _MklFusedInstanceNorm related configuration

* Enable QDQ-adapted _MklFusedInstanceNorm+[Relu/LeakyRelu] fusion and quantization

* Add attribute 'reduction_axes' to qin

* Add dummy mean/variance nodes for _MklFusedInstanceNorm

* Add QIN freeze_value

* Fix QIN freeze_value

* Add performance_only mode for _MklFusedInstanceNorm
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants