Abstract
We present an autoencoder that leverages learned representations to better measure similarities in data space. By combining a variational autoencoder (VAE) with a generative adversarial network (GAN) we can use learned feature representations in the GAN discriminator as basis for the VAE reconstruction objective. Thereby, we replace element-wise errors with feature-wise errors to better capture the data distribution while offering invariance towards e.g. translation. We apply our method to images of faces and show that it outperforms VAEs with element-wise similarity measures in terms of visual fidelity. Moreover, we show that the method learns an embedding in which high-level abstract visual features (e.g. wearing glasses) can be modified using simple arithmetic.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 33rd International Conference on Machine Learning (ICML 2016) |
| Number of pages | 9 |
| Publication date | 2016 |
| Publication status | Published - 2016 |
| Event | 33rd International Conference on Machine Learning (ICML 2016) - New York, United States Duration: 19 Jun 2016 → 24 Jun 2016 Conference number: 33 http://icml.cc/2016/ |
Conference
| Conference | 33rd International Conference on Machine Learning (ICML 2016) |
|---|---|
| Number | 33 |
| Country/Territory | United States |
| City | New York |
| Period | 19/06/2016 → 24/06/2016 |
| Internet address |
| Series | JMLR: Workshop and Conference Proceedings |
|---|---|
| Volume | 48 |
| ISSN | 1938-7288 |