Abstract

Flow-based generative models have become an important class of unsupervised learning approaches. In this work, we incorporate the key ideas of renormalization group (RG) and sparse prior distribution to design a hierarchical flow-based generative model, RG-Flow, which can separate information at different scales of images and extract disentangled representations at each scale. We demonstrate our method on synthetic multi-scale image datasets and the CelebA dataset, showing that the disentangled representations enable semantic manipulation and style mixing of the images at different scales. To visualize the latent representations, we introduce receptive fields for flow-based models and show that the receptive fields of RG-Flow are similar to those of convolutional neural networks. In addition, we replace the widely adopted isotropic Gaussian prior distribution by the sparse Laplacian distribution to further enhance the disentanglement of representations. From a theoretical perspective, our proposed method has O(logL) complexity for inpainting of an image with edge length L, compared to previous generative models with O(L2) complexity.

Details

Title
RG-Flow: a hierarchical and explainable flow model based on renormalization group and sparse prior
Author
Hong-Ye, Hu 1   VIAFID ORCID Logo  ; Wu, Dian 2   VIAFID ORCID Logo  ; Yi-Zhuang, You 1 ; Olshausen, Bruno 3 ; Chen, Yubei 3   VIAFID ORCID Logo 

 Department of Physics, University of California , San Diego, La Jolla, CA 92093, United States of America 
 Computational Quantum Science Laboratory, École Polytechnique Fédérale de Lausanne , CH-1015 Lausanne, Switzerland 
 Redwood Center, Berkeley AI Research, University of California, Berkeley , Berkeley, CA 94720, United States of America 
First page
035009
Publication year
2022
Publication date
Sep 2022
Publisher
IOP Publishing
e-ISSN
26322153
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2697710829
Copyright
© 2022 The Author(s). Published by IOP Publishing Ltd. This work is published under http://creativecommons.org/licenses/by/4.0 (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.