Conversation
* initial commit * bug fixes and adjusting conv inputs * separate forward function for Discriminator and Generator and disable Gen training for debugging * remove debugger statement * bug fix * detaching stuff before accumulating * refactor and add component as optional parameter * Add optimizer for and backprop against encoder * Add in README
* initial commit * bug fixes and adjusting conv inputs * separate forward function for Discriminator and Generator and disable Gen training for debugging * remove debugger statement * bug fix * detaching stuff before accumulating * refactor and add component as optional parameter * Add optimizer for and backprop against encoder * Add in README * more fixes to eval mode * create optimizers so that they can be saved * fix typo
Codecov Report
@@ Coverage Diff @@
## master #263 +/- ##
==========================================
- Coverage 82.53% 82.48% -0.05%
==========================================
Files 205 206 +1
Lines 15829 15848 +19
==========================================
+ Hits 13064 13072 +8
- Misses 2765 2776 +11
Continue to review full report at Codecov.
|
* initial commit * bug fixes and adjusting conv inputs * separate forward function for Discriminator and Generator and disable Gen training for debugging * remove debugger statement * bug fix * detaching stuff before accumulating * refactor and add component as optional parameter * Add optimizer for and backprop against encoder * Add in README * more fixes to eval mode * create optimizers so that they can be saved * fix typo * linting issues * add type annotation for encoder * fix linting * Isolate AE in training * works after changing the learning rate * remove debugger
|
Please merge |
|
You might also want to change the PR title; should reference #166 instead. |
| train_op_g.zero_grad() | ||
| step += 1 | ||
|
|
||
| vals_d = model(batch, gamma_, lambda_g_, mode="train", |
There was a problem hiding this comment.
Which dataset do you use here? train_g? It seems that train_d is used in texar-tf. Why does such a difference exist?
There was a problem hiding this comment.
For training, I use the same dataset because it is tricky to switch datasets between steps. And considering the discriminator and generator are trained separately, I've noticed it doesn't really affect the results.
There was a problem hiding this comment.
Removing the 2 iterators and keeping just one train iterator.
There was a problem hiding this comment.
If that is the case, we do not need to have train_d in the iterator (?)
| hparams=config.model['opt'] | ||
| ) | ||
|
|
||
| def _train_epoch(gamma_, lambda_g_, epoch, verbose=True): |
There was a problem hiding this comment.
In texar-tf, both train_d and train_g are used in the _train_epoch, but we only use train_g in this function. Did I miss something important?
* Reviewed changes * linting
* initial commit * linting
initial commit
bug fixes and adjusting conv inputs
separate forward function for Discriminator and Generator and disable Gen training for debugging
remove debugger statement
texar bug fix
detaching stuff before accumulating
refactor and add component as optional parameter
Add optimizer for and backprop against encoder
Add in README