Restricted Boltzmann Machine

Restricted Boltzmann Machine (RBM) is a two-layered neural network--the first layer is called the visible layer and the second layer is called the hidden layer. They are called shallow neural networks because they are only two layers deep. They were first proposed in 1986 by Paul Smolensky (he called them Harmony Networks[1]) and later by Geoffrey Hinton who in 2006 proposed Contrastive Divergence (CD) as a method to train them. All neurons in the visible layer are connected to all the neurons in the hidden layer, but there is a restriction--no neuron in the same layer can be connected. All neurons are binary in nature:

Source: By Qwertyus - Own work, CC BY-SA 3.0, https://commons.wikimedia.org/w/index.php?curid=22717044

RBM can be used for dimensionality reduction, feature extraction, and collaborative filtering. The training in RBMs can be divided into three parts: forward pass, backward pass, and then compare.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.145.76.250