-
Notifications
You must be signed in to change notification settings - Fork 309
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how to express 3 discrete latent codes (each with dimension 20) and visual work ok? #10
Comments
if isinstance(dist, Gaussian): |
https://github.com/RutgersHan/InfoGAN/blob/dev_auto/launchers/generate_images.py |
C.3 CelebA is
but how to config C.5 Chairs as below The network architectures are shown in Table 6. The discriminator D and the recognition network Q shares the same network, and only have separate output units at the last layer. For this task, we use 1 continuous latent code, 3 discrete latent codes (each with dimension 20), and 128 noise variables, so the input to the generator has dimension 189. elif isinstance(dist, Bernoulli): |
The above latent_spec worked okay for me. c3_celebA_latent_spec = [
(Uniform(128), False), # Noise
(Categorical(10), True),
(Categorical(10), True),
(Categorical(10), True),
(Categorical(10), True),
(Categorical(10), True),
(Categorical(10), True),
(Categorical(10), True),
(Categorical(10), True),
(Categorical(10), True),
(Categorical(10), True),
]
c3_celebA_image_size = 32 Can you elaborate a bit more in words what you're having problems with? I'm not sure I understand what's not working for you. |
I might be misunderstanding, but it seems like
would translate to the following latent_spec. That is, the continuous code is represented by c5_chairs_latent_spec = [
(Uniform(128), False), # Noise
(Uniform(1, fix_std=True), True),
(Categorical(20), True),
(Categorical(20), True),
(Categorical(20), True),
]
c3_celebA_image_size = 32 I copied the I'm not sure where you got the LatentGaussian from... I don't know if it's necessary? I haven't tried running the Chairs model at all. |
This is correct. Thanks @NHDaly !
Yes, to better compare with previous supervised results, we select codes from multiple runs that are most similar to categories that previous method (DC-IGN) produces. |
for @NHDaly ref this https://github.com/RutgersHan/InfoGAN/blob/dev_auto/launchers/run_flower_exp.py#L49 is your celeba train result is ok? for @neocxi 1 2 3 what cause NAN error? D and G learning rate not equilibrium?? Epoch 14 | discriminator_loss: 0.128064; generator_loss: 2.78964; MI_disc: 20.3559; CrossEnt_disc: 2.66993; MI: 20.3559; CrossEnt: 2.66993; max_real_d: 0.999938; min_real_d: 0.560705; max_fake_d: 0.240968; min_fake_d: 0.0144349 4 celeba train need how long ? epoch log can share ? log d loss very small g loss bigger |
A.2 INFOGAN TRAINING from |
how to express 10 dimensional categorical variables
code:
latent_spec = [
(Uniform(62), False),
(Categorical(10), True),
(Uniform(1, fix_std=True), True),
(Uniform(1, fix_std=True), True),
]
is for mnist ,
this is not enough,
in paper:
MNIST, we choose to model the latent codes with one categorical code, c1 ⇠ Cat(K = 10, p = 0.1), which can model discontinuous variation in data, and two continuous codes that can capture variations that are continuous in nature: c2 , c3 ⇠ Unif ( 1, 1).
but what to express: Street View House Number (SVHN
we make use of four 10 dimensional categorical variables and two uniform continuous variables as latent codes.
CelebA
In this dataset, we model the latent variation as 10 uniform categorical variables, each of dimension 10.
append c.3
generator G
Input 2 R228 228 how to get 228?
discriminator D / recognition network Q generator G
Input 32 ⇥ 32 Color image Input 2 R228
4 ⇥ 4 conv. 64 lRELU. stride 2 FC. 2 ⇥ 2 ⇥ 448 RELU. batchnorm
4 ⇥ 4 conv. 128 lRELU. stride 2. batchnorm 4 ⇥ 4 upconv. 256 RELU. stride 2. batchnorm 4 ⇥ 4 conv. 256 lRELU. stride 2. batchnorm 4 ⇥ 4 upconv. 128 RELU. stride 2.
FC. output layer for D,
FC.128-batchnorm-lRELU-FC.output for Q 4 ⇥ 4 upconv. 64 RELU. stride 2.
4 ⇥ 4 upconv. 3 Tanh. stride 2.
any one any help?
thanks very much !
The text was updated successfully, but these errors were encountered: