-
Notifications
You must be signed in to change notification settings - Fork 2
Deval's daily progress report
Deval Srivastava edited this page Jun 19, 2018
·
5 revisions
- Researched about types of GANs
- Looked up ways to generate images from some text by creating a text encoding
- Researched ways to generate video from images
- Found a way to use an LSTM network to generate video from images
- Generating videos from images by separating the static background from foreground
- Referred some PyTorch tutorials (refer)
- Started to work on the report as to how to tackle the problem, various algorithms available
- Completed the task 1 report
- Read about temporal conditioned GANs to generate video from text
- Discussed Task 2 details
- Went through a few GAN Tutorials
- Went through a few pyTorch Tutorials
- Looked up DCGAN architecture details
- Programmed the code for DCGAN
- Trained the DCGAN model
- committed task 2 with comments
- discussed task 2 details
- researched more about text-to-image
- started work on creating the model
- found out about stack GAN
- looked up datasets for training our model
- still working on creating the model
- looking up datasets for text captions
- started work on the progress presentation
- dataset generation script was prepared
- finalized the ppt for now
- removed last of the errors from the model
- put the model on training.
- progress presentation performed.
- found out that the training was unsuccessful as the images were very poor quality and the loss curve was flat lining
- made some improvements to the model.
- the training images were yet the same and the loss curve continued to flat line
- decided to change the model and use stackGAN.
- the training had completed and some images were generated from the coco dataset.
- looking to find appropriate dataset for video generation. -- http://www.nada.kth.se/cvap/actions/ -- http://crcv.ucf.edu/data/UCF101.php -- http://www.cs.toronto.edu/~nitish/unsupervised_video/
- as per kalind sir's suggestion used the birds/flowers dataset for image generation.
- prepared the dataloader for birds dataset. -- http://www.vision.caltech.edu/visipedia/CUB-200-2011.html -- http://www.robots.ox.ac.uk/~vgg/data/flowers/102/
- put the flowers model for training.
- continued work for video dataset and dataloader.
- put the birds model on training