Skip to content

MuserHao/Kernel-stein-discrepancy-for-energy-based-model

Repository files navigation

Kernel-stein-discrepancy-for-energy-based-model

Key papers on Score-based Generative Modeling:

Paper Description
Generative Modeling by Estimating Gradients of the Data Distribution First score-based generative modeling: Denoising Score Matching + Annealed Langevin Dynamics
How to Train Your Energy-Based Models A nice tutorial/introduction/summary on training EBM by Yang Song, including maximum likelihood, score matching, contrastive divergence and adversarial training.
Denoising Diffusion Probabilistic Models A diffusion probabilistic models training on a weighted variational bound designed according to a novel connection between diffusion probabilistic models and denoising score matching with Langevin dynamics.
Score-Based Generative Modeling through Stochastic Differential Equations A generative model framework based on SDE that encapsulates previous approaches in score-based generative modeling and diffusion probabilistic modeling, allowing for new sampling procedures and new modeling capabilities. In particular, we introduce a predictor-corrector framework to correct errors in the evolution of the discretized reverse-time SDE.
On Maximum Likelihood Training of Score-Based Generative Models In this note, we show that such an objective is equivalent to maximum likelihood for certain choices of mixture weighting.
Neural Ordinary Differential Equations We introduce a new family of deep neural network models. Instead of specifying a discrete sequence of hidden layers, we parameterize the derivative of the hidden state using a neural network. The output of the network is computed using a black-box differential equation solver.
Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting Proposed the convolutional LSTM (ConvLSTM) and use it to build an end-to-end trainable model for the precipitation nowcasting problem.
Deep Learning for Precipitation Nowcasting: A Benchmark and A New Model Proposed trajectory GRU (TrajGRU) model that can actively learn the location-variant structure for recurrent connections. With the goal of making high-resolution forecasts.

Kernelized Stein's Discrepancy:

Paper Description
A Kernelized Stein Discrepancy for Goodness-of-fit Tests and Model Evaluation Kernelized Stein Discrepancy (KSD) as a computable measure of discrepancy between a sample of an unnormalized distribution
Kernelized Complete Conditional Stein Discrepancy Conditional kernelized Stein discrepancy using joint kernel
Learning the Stein Discrepancy for Training and Evaluating Energy-Based Models without Sampling Learning Stein discrepancy optimal evaluation function using neural network
Minimum Stein Discrepancy Estimators Provide a unifying perspective of these techniques as minimum Stein discrepancy estimators, and use this lens to design new diffusion kernel Stein discrepancy (DKSD) and diffusion score matching (DSM) estimators with complementary strengths.
Derivative reproducing properties for kernel methods in learning theory Deriavtives of functions in RKHS to hold reproducing properties

Learning Energy based model (EBM):

Paper Description
EBMs Trained with Maximum Likelihood are Generator Models Trained with a Self-adverserial Loss The authors show that reintroducing the noise in the dynamics does not lead to a qualitative change in the behavior, and merely reduces the quality of the generator. We thus show that EBM training is effectively a self-adversarial procedure rather than maximum likelihood estimation.
On Maximum Likelihood Training of Score-Based Generative Models In this note, we show that such an objective is equivalent to maximum likelihood for certain choices of mixture weighting.

Kernel Mean Embedding and Conditional Mean Embedding:

Paper Description
Kernel Mean Embedding of Distributions: A Review and Beyond A review of kernel mean embedding and conditional kernel mean embedding for distribution
Kernel dimension reduction in regression Kernel dimension reduction
Kernel Conditional Density Operators Conditional Density estimation with kernel mean embedding method
McGan: Mean and Covariance Feature Matching GAN Training Generative Adversarial Networks (GAN) based on matching statistics of distributions embedded in a finite dimensional feature space. Mean and covariance feature matching IPMs.

Nonparametric Conditional Density Estimation

Paper Description
Nonparametric Density Estimation for High-Dimensional Data - Algorithms and Applications A review paper on nonparametric density estimation for high-dimensional data
Conditional Density Estimation with Neural Networks: Best Practices and Benchmarks The paper develops best practices for conditional density estimation for finance applications with neural networks, grounded on mathematical insights and empirical evaluations.
Nonparametric Conditional Density Estimation in a High-Dimensional Regression Setting Proposed a new nonparametric estimator of conditional density that adapts to sparse (low-dimensional) structure in x.

Conditional_VAE

Related papers:

Paper Description
Learning Structured Output Representation using Deep Conditional Generative Models Conditional VAE
Deep Generative Models with Learnable Knowledge Constraints Learn generative models under constraints
Semi-Supervised Learning with Deep Generative Models VAE as semi-supervised learning
Projection Pursuit Regression Projection pursuit regression
Variational Autoencoder with Arbitrary Conditioning Variational autoencoder that can be conditioned on an arbitrary subset of observed features and then sample the remaining features in "one shot".
Infinite Variational Autoencoder for Semi-Supervised Learning Train a generative model using unlabelled data, and then use this model combined with labelled data to train a discriminative model for classification.
Adversarial Autoencoders "adversarial autoencoder" (AAE), which is a probabilistic autoencoder that uses the recently proposed generative adversarial networks (GAN) to perform variational inference
Adversarial Symmetric Variational Autoencoder Adversarial Symmetric Variational Autoencoder, using a adversial training in VAE, which has a very interesting comment on the Maximum likelihood estimation

Semi-supervised learning with Deep Generative Models: Pytorch implementation

Generative models (AAE included): Pytorch implementation

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages