An Interpretable Generative Model for Handwritten Digits Synthesis
An interpretable generative model for handwritten digits synthesis is proposed in this work. Modern image generative models such as the variational autoencoder (VAE) are trained by backpropagation (BP). The training process is complex, and its underlying mechanism is not transparent. Here, we presen...
Saved in:
Published in | Proceedings - International Conference on Image Processing pp. 1910 - 1914 |
---|---|
Main Authors | , , , , , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
01.09.2019
|
Subjects | |
Online Access | Get full text |
ISSN | 2381-8549 |
DOI | 10.1109/ICIP.2019.8803129 |
Cover
Summary: | An interpretable generative model for handwritten digits synthesis is proposed in this work. Modern image generative models such as the variational autoencoder (VAE) are trained by backpropagation (BP). The training process is complex, and its underlying mechanism is not transparent. Here, we present an explainable generative model using a feedforward design methodology without BP. Being similar to VAEs, it has an encoder and a decoder. For the encoder design, we derive principal-component-analysis-based (PCA-based) transform kernels using the covariance of its inputs. This process converts input images of correlated pixels to uncorrelated spectral components, which play the same role as latent variables in a VAE system. For the decoder design, we convert randomly generated spectral components to synthesized images through the inverse PCA transform. A subject test is conducted to compare the quality of digits generated using the proposed method and the VAE method. They offer comparable perceptual quality yet our model can be obtained at much lower complexity. |
---|---|
ISSN: | 2381-8549 |
DOI: | 10.1109/ICIP.2019.8803129 |