We present a theoretical analysis of Gaussian-binary restricted Boltzmann machines (GRBMs) from the perspective of density models. The key aspect of this analysis is to show that GRBMs can be formulated as a constrained mixture of Gaussians, which gives a much better insight into the model's capabilities and limitations. We further show that GRBMs are capable of learning meaningful features without using a regularization term and that the results are comparable to those of independent component analysis. This is illustrated for both a two-dimensional blind source separation task and for modeling natural image patches. Our findings exemplify that reported difficulties in training GRBMs are due to the failure of the training algorithm rather than the model itself. Based on our analysis we derive a better training setup and show empirically that it leads to faster and more robust training of GRBMs. Finally, we compare different sampling algorithms for training GRBMs and show that Contrastive Divergence performs better than training methods that use a persistent Markov chain.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC5289828PMC
http://journals.plos.org/plosone/article?id=10.1371/journal.pone.0171015PLOS

Publication Analysis

Top Keywords

training grbms
12
gaussian-binary restricted
8
restricted boltzmann
8
boltzmann machines
8
modeling natural
8
natural image
8
better training
8
grbms
6
training
6
machines modeling
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!