20 Epochs on COCO - (Larger Transformer) #335
Replies: 5 comments 9 replies
-
I'm curious of the impact to use these axial row and axial col layers vs full attn. I might try it too |
Beta Was this translation helpful? Give feedback.
-
nice results! |
Beta Was this translation helpful? Give feedback.
-
Nice results! |
Beta Was this translation helpful? Give feedback.
-
@afiaka87 |
Beta Was this translation helpful? Give feedback.
-
This one was done entirely on one GPU over the course of about 48 hours. It has no dependencies on DeepSpeed and was trained with 32 bit precision.
If you want to train your own DALL-E and have about 8 GiB of VRAM; this is a very good configuration in my opinion.
Full W&B training session
Run inference with the following notebook:
https://colab.research.google.com/drive/1UqTIilYvILiO_mH88LP67YkBRmc0dEFR
Details
Transformer:
VQGAN:
Training
Augmentations:
left,right,top,bottom
to describe where things are on the image.Hardware:
Checkpoints:
You can find a checkpoint for each epoch trained here:
https://wandb.ai/dalle-pytorch-replicate/COCO512_16_16D_16H_80TSL/artifacts/model/trained-dalle/07c445559fd9183e302e
Beta Was this translation helpful? Give feedback.
All reactions