PixelVAE icon indicating copy to clipboard operation
PixelVAE copied to clipboard

one-level network for 64x64 image size datasets

Open jcpeterson opened this issue 7 years ago • 11 comments

Any chance you can add this?

jcpeterson avatar Jun 25 '17 22:06 jcpeterson

I can put something together, but we ran experiments with this and found it didn't perform nearly as well as 2-level. Do you want it anyway?

igul222 avatar Jun 26 '17 22:06 igul222

Yes, that would be much appreciated. I'd like to compare performance on a few datasets under several conditions, especially celebA, which I can already successfully load in as jpg files. Thanks

jcpeterson avatar Jun 26 '17 22:06 jcpeterson

It doesn't have to run with this codebase if you have another that works already

jcpeterson avatar Jun 29 '17 07:06 jcpeterson

Just updated the code with a model that should work ('64px_big_onelevel'), but I haven't actually tested it. Let me know how it goes if you do.

igul222 avatar Jun 29 '17 19:06 igul222

Thanks Ishaan! Likely citation incoming ;)

I will give this a shot and let you know if it runs / how it performs.

I don't know if you are aware, but this architecture seems quite SOTA for smallish datasets with manageable variation (more complex than MNIST, but less than CIFAR where data-to-complexity ratio seems off to me), outperforming even the best new GANs. That doesn't really come across in the paper.

jcpeterson avatar Jun 29 '17 20:06 jcpeterson

Interesting! SOTA in what sense, sample quality? I'd say LSUN bedrooms fits the description you mentioned (plenty of data, not terribly complex images) and we tried to demonstrate pretty good results there I think.


From: Joshua Peterson [email protected] Sent: Thursday, June 29, 2017 1:24:10 PM To: igul222/PixelVAE Cc: Ishaan Gulrajani; Comment Subject: Re: [igul222/PixelVAE] one-level network for 64x64 image size datasets (#2)

Thanks Ishaan! Likely citation incoming ;)

I will give this a shot and let you know if it runs / how it performs.

I don't know if you are aware, but this architecture seems quite SOTA for smallish datasets with manageable variation (more complex than MNIST, but less than CIFAR where data-to-complexity ratio seems off to me), outperforming even the best new GANs. That doesn't really come across in the paper.

— You are receiving this because you commented. Reply to this email directly, view it on GitHubhttps://github.com/igul222/PixelVAE/issues/2#issuecomment-312093477, or mute the threadhttps://github.com/notifications/unsubscribe-auth/ABBP7nyXlCuQyopMjToM5B_yZ5w2UdH7ks5sJAfqgaJpZM4OEvS3.

igul222 avatar Jun 29 '17 20:06 igul222

Sample quality, but also a lower probability of artifacts that give away that the images aren't "real". Agreed on LSUN, and clearly you tried a large range of datasets, much more than most papers.

I can show you some examples soon enough.

jcpeterson avatar Jun 29 '17 21:06 jcpeterson

Ok, it looks like training works with the new script, thanks! The sampling code does seem different though. The samples are int32 (isn't this wrong?), and it samples 8 points and then adds variability to those?

jcpeterson avatar Jul 08 '17 16:07 jcpeterson

Actually, the int32 is for the output images it seems, so that's fine. However, I'm still not sure where the variability is coming from in the rows of the output image

jcpeterson avatar Jul 08 '17 18:07 jcpeterson

Can you point to the sampling code you're talking about? The code that gets executed should be lines 855-902.

igul222 avatar Jul 08 '17 20:07 igul222

Yes that's right. When this runs, I get 8 samples (rows), and then 8 variations (columns) on those samples. I'd like sampling behavior like the two-level network with the logits *= 10000. line added. The redundant columns are easy to remove, so that's not really a problem.

jcpeterson avatar Jul 18 '17 00:07 jcpeterson