Seems like a promising algorithm to create compressed latent spaces. The basic idea is that you jointly train some loss functions that learn to estimate the amount of mutual information between the latent space and the image. Very cool!
Thank you for sharing this! (And kudos for giving credit to the chainer implementation.)
I know I always like the fun pictures best when implementing things, but did you, by any chance, try to reproduce some of their evaluation metric results, e.g. Table 1 and 2 on page seven? (Again, if you didn’t, it’s totally cool, and I’m grateful that you share what you have.)
Just an update on this, so I ran the classification test on the 64 bit latent space. In the paper that’s the result on Table 1, row DIM(L) (NCE) , column CIFAR 10, Y(64)
I’ll put the results up, but I got a classification accuracy of 60.43, whereas they report 69.13 in the paper.
I’m not able to account for the difference as of yet, as they have not released their code. So it’s possible they are doing something different than what I’m doing in the model.
Having said that, even 60 is a better result than what they report for the other encoding methods. So it’s not all bad news.
Hi, we’ve had an implementation up for a while now:
Though the NCE version with the “deep bilinear” model isn’t up yet. We are in the process of updating the repo.
That said, 60% is quite low and even with JSD and the concat-style scoring function should get at least 64-65% with the global vector (iirc) and over 71% for the local feature maps. Give our current repo a try (and the updated version once its finished), and let me know if you are encountering the same problems.