Skip to content

jerryfung110/Cluster_Algorithm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

1 Commit
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Semisupervised-VAE-for-Regression-Application-on-Soft-Sensor

We present the development of a semi-supervised regression method using variational autoencoders (VAE), which is customized for use in soft sensing applications. We motivate the use of semi-supervised learning considering the fact that process quality variables are not collected at the same frequency as other process variables leading to many unlabelled records in operational datasets. These unlabelled records are not possible to use for training quality variable predictions based on supervised learning methods. Use of VAEs for unsupervised learning is well established and recently they were used for regression applications based on variational inference procedures. We extend this approach of supervised VAEs for regression (SVAER) to make it learn from unlabelled data leading to semi-supervised VAEs for regression (SSVAER), then we make further modifications to their architecture using additional regularization components to make SSVAER well suited for learning from both labelled and unlabelled process data. The probabilistic regressor resulting from the variational approach makes it possible to estimate the variance of the predictions simultaneously, which provides an uncertainty quantification along with the generated predictions. We provide an extensive comparative study of SSVAER with other publicly available semi-supervised and supervised learning methods on two benchmark problems using fixed-size datasets, where we vary the percentage of labelled data available for training. In these experiments, SSVAER achieves the lowest test errors in 11 of the 20 studied cases, compared to other methods where the second best gets 4 lowest test errors out of the 20.

SSVAER scheme{:width="200px"}

Results

Compare on debutanizer dataset

Label% 1% 2% 5% 10% 14.2% 20% 25% 33% 50% 100%
SSVAER 0.0764 0.0522 0.0561 0.0498 0.0476 0.0470 0.0480 0.0476 0.0469 0.0516
SVAER 0.0606 0.0588 0.0541 0.0478 0.0507 0.0589 0.0545 0.0490 0.0540 0.0543
SSAE 0.0822 0.0526 0.0494 0.0495 0.0485 0.0529 0.0496 0.0512 0.0521 0.0462
FCNN 0.0852 0.0541 0.0588 0.0517 0.0483 0.0487 0.0487 0.0493 0.0493 0.0495
SSSMM NA NA NA NA NA 0.0525 0.0731 0.0629 0.0647 NA

Compare on sulfur recovery unit dataset

Label% 1% 2% 5% 10% 14.2% 20% 25% 33% 50% 100%
SSVAER 0.0566 0.0484 0.0347 0.0322 0.0289 0.0314 0.0275 0.0285 0.0268 0.0290
SVAER 0.0568 0.0506 0.0354 0.0322 0.0275 0.0317 0.0282 0.0320 0.0302 0.0315
SSAE 0.0603 0.0499 0.0410 0.0347 0.0308 0.0358 0.0287 0.0254 0.0274 0.0270
FCNN 0.0653 0.0459 0.0461 0.0342 0.0330 0.0362 0.0341 0.0298 0.0367 0.0382
SSSMM NA NA NA 0.0580 NA 0.0351 0.0318 0.0560 0.0445 NA

Line plots of the two benchmark datasets

Variations of RMSEs against percentage of labelled entries on the two test datasets{:width="200px"}

Confidence interval plot

By utilizing the estimated variance along with the quality variable prediction, the confidence interval can be shown as:

95% confidence interval plot on debutanizer test data (20%label){:width="200px"}

95% confidence interval plot on debutanizer test data (20%label){:width="200px"}

Demos:

demo_debut.ipynb

demo_sru.ipynb

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 94.3%
  • Python 5.7%