Understanding the math behind the neural network

The neurons present in different layers--input, hidden, and output--are interconnected through a mathematical function called activation function, as displayed in Figure 1. There are different variants of the activation function, which are explained as follows. Understanding the activation function will help in implementation of the neural network model for better accuracy:

  • Sigmoid function: This is frequently used by professionals in data mining and analytics, as it is easier to explain and implement too. The equation is mentioned here:

    Understanding the math behind the neural network

    The sigmoid function, also known as the logistic function, is mostly used to transform the input data from the input layer to the mapping layer, or the hidden layer.

  • Linear function: This is one of the simple functions typically used to transfer information from the de-mapping layer to the output layer. The formula is as follows:

                                                                           f(x)=x

  • Gaussian function: Gaussian functions are bell-shaped curves that are applicable for continuous variables, where the objective is to classify the output into multiple classes:

    Understanding the math behind the neural network

  • Hyperbolic tangent function: This is another variant of transformation function; it is used to transform information from the mapping layer to the hidden layer:

    Understanding the math behind the neural network

  • Log sigmoid transfer function: The following formula explains the log sigmoid transfer function used in mapping the input layer to the hidden layer:

    Understanding the math behind the neural network

  • Radial basis function: This is another activation function; it is used to transfer information from the de-mapping layer to the output layer:

    Understanding the math behind the neural network

Different types of transfer functions, as previously discussed, can be interchangeable in neural network architectures. They can be used in different stages such as input to hidden, hidden to output, and so on, to improve the model accuracy.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.139.82.4