Skip to content

Conversation

@echarlaix
Copy link
Collaborator

No description provided.

@echarlaix echarlaix merged commit a38b66b into main Jun 6, 2022
@echarlaix echarlaix deleted the add-logo branch June 6, 2022 06:46
PenghuiCheng pushed a commit that referenced this pull request Sep 28, 2022
yujiepan-work referenced this pull request in yujiepan-work/optimum-intel Dec 22, 2022
#2)

* use nncf official branch for install since JPQD is merged

* copy ac scripts from transformer repo

* init commit for wav2vec2

* add onnx_config argument in OVTrainer for onnx export with unsupported model

* enable customized teancher kd

* add readme

* delete debugging lines
yujiepan-work referenced this pull request in yujiepan-work/optimum-intel Jan 8, 2023
#2)

* use nncf official branch for install since JPQD is merged

* copy ac scripts from transformer repo

* init commit for wav2vec2

* add onnx_config argument in OVTrainer for onnx export with unsupported model

* enable customized teancher kd

* add readme

* delete debugging lines
yujiepan-work referenced this pull request in yujiepan-work/optimum-intel Jan 30, 2023
#2)

* use nncf official branch for install since JPQD is merged

* copy ac scripts from transformer repo

* init commit for wav2vec2

* add onnx_config argument in OVTrainer for onnx export with unsupported model

* enable customized teancher kd

* add readme

* delete debugging lines
yujiepan-work referenced this pull request in yujiepan-work/optimum-intel Feb 8, 2023
#2)

* use nncf official branch for install since JPQD is merged

* copy ac scripts from transformer repo

* init commit for wav2vec2

* add onnx_config argument in OVTrainer for onnx export with unsupported model

* enable customized teancher kd

* add readme

* delete debugging lines
yujiepan-work referenced this pull request in yujiepan-work/optimum-intel Feb 11, 2023
#2)

* use nncf official branch for install since JPQD is merged

* copy ac scripts from transformer repo

* init commit for wav2vec2

* add onnx_config argument in OVTrainer for onnx export with unsupported model

* enable customized teancher kd

* add readme

* delete debugging lines
echarlaix pushed a commit that referenced this pull request Feb 21, 2023
)

* Initial commit to enable OVTrainer with joint pruning, quantization and distillation via NNCF

* Review OpenVINO Q&A readme and configs

* Update README.md

* Add post init value checker to OVTrainingArguments

* Initial enabling of audio classification/wav2vec2 [tests not included] (#2)

* use nncf official branch for install since JPQD is merged

* copy ac scripts from transformer repo

* init commit for wav2vec2

* add onnx_config argument in OVTrainer for onnx export with unsupported model

* enable customized teancher kd

* add readme

* delete debugging lines

* Update openvino-dev and nncf version in setup.py

* refactor _enable_standard_onnx_export_option to _set_standard_onnx_export_option

* add tests for (movement/quantization) with distillation (#3)

* test part 1

* clean "compute_distillation_loss' in OVTrainer

* add test of OVTrainer for int8+kd / movement / movement+int8/ movement+int8+kd

* add expectedFailuremark to test of OVModelForAudioClassification

* revert unncessary codes about "OVModelForAudioClassification"

* change to a shorter train for w2v2 in readme

* revert compute_metrics change since it is not unnecessary

* fix task_loss non-scalar bug for kd logging

* make regex clearer in QA bert config

* Refactor compression-related logging

* Refactor OpenVINO IR generation and patch tests

* Miscellaneous refactoring
* MO IR pruning depends on scheduler stage
* Readme tweaks for all example tasks
* Minor tweak on tests
* Align setup.py for openvino-dev and nncf versions needed for JPQD

* Fix lint with Black

* Refactor OpenVINO IR generation using python api

* Fix via isort

* Handle IR generation error to avoid run termination

* Update QA readme

* Enable distillation on openvino's image classification example

* Minor refactoring in openvino's audio classification example

* Move openvino-dev dependency to be extra of NNCF

* Configure IR model to accept dynamic-shaped input

* Revert _enable_standard_onnx_export_option method in OVConfig

* Update wav2vec2 configs for audio classification

* Add BERT-base/glue-sst2 example with QAT /  JPQD (#4)

* copy text-classification example from transformers

* init draft for sst example

* update sst2 accuracy & training time

* Revise wav2vec2 config and audio classification readme

* Patch _enable_standard_onnx_export_option to only add the key pair to quantization config

* Set logging level to INFO in openvino/trainer.py

* Review readme of text and image classification

* Revert IR generation with static input shape for joint compression

* Add distillation and advance optimization section in optimization_ov.mdx

* Patch tests

* Revise formatting of optimization_ov.mdx

* Limit #checkpoint saved for JPQD samples

* Handle NNCF output to text log and only print errors to stdout

* Replace hardcoded model.onnx filename with constant variable

* Fix movement sparsity config in optimization_ov.mdx

* Change _set_feature to _set_task to align with OVQuantizer

* Revert onnx_config exposure in OVTrainer, expand test coverages for joint compression variations, misc. patches

* use builtin onnx configs for wav2vec onnx export

* move teacher model argument from OVTrainingArgs to model args

* fix duplicate call of `epoch_step`

* temporal workaround about compression metrics

* test for all training cases

* temporal workaround for eval only

* cover train/eval tests

* style fix

* Move old ovtrainer tests to a new `test_training.py` file;  bugfix in training loss check (#6)

* removing old tests in test_quantization since they are now in `test_training`

* bugfix in checking compression metrics during training

* keep bert examples only and misc. fixes (#7)

* temporarily keep bert examples only; remove w2v2 and swin

* move nncf_compression_config out of OVTrainingArguments

* type hint change for nncf_compression_config

* documnet rename feature to task

* revert existing QAT image classification example

* delete useless codes in test quantization

* revert existing test_ quantization

* misc change in compute_metric

* revert unnecessary changes

* temporal workaround for logging distill & compression loss (not using dist. reduce)

* revert set_task method

* bugfix in compression metric in qa task

* bugfix in importing tpu

* simplify pruning ir codes

* clean unncessary distillation weight attribute in trainer

* Change nncf requirement to official 2.4

* Log nncf compression statistics at the beginning of each training epoch

* Revise optimization_ov.mdx documentation
* Consolidate during training optimization to QAT and JPQD
* Add known limitation regarding OpenVINO IR with static input shape

* fix data parallel crashes and add tests for DP/DDP (#8)

* fix "not same device"  error in data parallel

* wrap teacher model with data parallel

* add sst2 tests for dp/ddp with fixes

* Add remark in optimization_ov.mdx on supported model architecture for structured pruning

* Refactor JPQD IR generation where final IR is dynamic in input shape

* Revise optimization_ov.mdx to remove static IR limitations
* revert snippet for inference with Transformers pipeline

* Remove commented codes in openvino/trainer.py

* Add tests about new OV IR export - check dynamic graph and output equivalence to torch model  (#9)

* draft for new export with some todos

* draft for tests

* delete onnx export debugging when errors on saving

* add back the debug info when ir export fails

* bugfix in random setting zeros in movement masks

* Add tests on OV IR reshape-ability

* Remove unused imports in openvino/trainer.py

* Refine inference pipeline with OVModel in optimization_ov.mdx

* Revise openvino extras in setup.py

---------

Co-authored-by: Alexander Kozlov <alexander.kozlov@intel.com>
Co-authored-by: Yujie Pan <yujie.pan@intel.com>
AlexKoff88 pushed a commit that referenced this pull request Dec 23, 2024
* Support AWQ models

* Add tests

* Add dependencies

* Fix tests

* enable awq export only if ov support it

* fix style (#2)

* disable awq and gptq install for old torch (#3)

* fix style

* disable autogptq and autoawq install for old transformers testing

* separate common quant models patching and gptq (#4)

* disable windows install (#5)

* separate common quant models patching and gptq

* disable awq windows

* skip logits check for quantized models (#6)

* fix test after rebase

* fix testing condition for 2024.6 and unpatch in case if failed

* Fix qwen2-vl tests (#1084)

* Skip private mdoel loading test for external contributors (#1082)

* Fix reshaping unet if timestep is 0d tensor (#1083)

* Disable kv cache compression for fp vlm (#1080)

* Support AWQ models

* Add tests

* Add dependencies

* Fix tests

* enable awq export only if ov support it

* fix style (#2)

* disable awq and gptq install for old torch (#3)

* fix style

* disable autogptq and autoawq install for old transformers testing

* separate common quant models patching and gptq (#4)

* disable windows install (#5)

* separate common quant models patching and gptq

* disable awq windows

* skip logits check for quantized models (#6)

* fix test after rebase

* fix testing condition for 2024.6 and unpatch in case if failed

* add necessary packages in test_openvino_full

* fix code style after rebase (#7)

---------

Co-authored-by: eaidova <ekaterina.aidova@intel.com>
Co-authored-by: Nikita Savelyev <nikita.savelyev@intel.com>
Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
ljaljushkin added a commit to ljaljushkin/optimum-intel that referenced this pull request Nov 14, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants