An ANN (Artificial Neural Network) is based on a collection of connected units or nodes called artificial neurons, which loosely model the neurons in a biological brain.
- Optimize memory manipulation in matrix operations.
- CSV or YAML input files for training data.
- Exportable Neuron Data.
- AUC (Area Under the Curve) and Accuracy metrics.
Figure. 1. Simple Neural Network Structure.
Based on nature, neural networks are the usual representation we make of the brain: Neurons interconnected to other neurons which forms a network. The AN's are composed of various layers containing n number of neurons.
Neural networks can usually be read from left to right. Here, the first layer is the layer in which inputs are entered. There are n amount of internals layers (called hidden layers) that do some math (in this case ther is only one), and one last layer that contains all the possible outputs.
Figure. 2. Neurons inner operations.
First, the nueron adds up the value of every neurons from the previous column it is connected to. On the Figure 2, there are 3 inputs (x1, x2, x3) coming to the neuron, so 3 neurons of the previous column are connected to our neuron.
This value is multiplied, before being added, by another variable called “weight” (w1, w2, w3) which determines the connection between the two neurons. Each connection of neurons has its own weight, and those are the only values that will be modified during the learning process.
Moreover, a bias value may be added to the total value calculated. It is not a value coming from a specific neuron and is chosen before the learning phase, but can be useful for the network.
After all those summations, the neuron finally applies a function called “activation function” to the obtained value.
Figure. 3. Sigmoid activation function.
The so-called activation function usually serves to turn the total value calculated before to a number between 0 and 1 (done for example by a sigmoid function shown by Figure 3). Other function exist and may change the limits of our function, but keeps the same aim of limiting the value.
Figure. 4. Neuron Operations.
The product of the input neurons and the weights can be represented by a matrix. To calculate the c[i,j] entry of the matrix C:=AB, one takes the dot product of the i'th row of the matrix A with the j'th column of the matrix B. For the dot product to be succesfuly performed, the rows of matrix A and columns of matrix B must have the same length.
| (W11 * I1) + (W21 * i2) + ... + (Wi1 * Ii) |
| (W12 * I1) + (W22 * i2) + ... + (Wi2 * Ii) |
C = | (W13 * I1) + (W23 * i2) + ... + (Wi3 * Ii) |
| · + ... + |
| (W1j + I1) + (W2j * I2) + ... + (Wij + Ii) |
Being the code equivalent in this library as:
// Being T a generic Type. (Float, Double, Int ...)
Matrix<T>* hidden_input_weights = Matrix<T>::dot(this->inputHiddenWeights, inputs);
Now, to add the bias:
hidden_input_weights->add(this->bias_hidden);
And finally pass it through the activation function.
this->hidden_weights_output = Matrix<T>::map(hidden_input_weights, sigmoid);
References: [1]: https://towardsdatascience.com/first-neural-network-for-beginners-explained-with-code-4cfd37e06eaf
Helpful statements: //"program": "${workspaceRoot}/out/build/${command:cmake.buildType}/${env:OS}/NeuralNetworkd", //"preLaunchTask": "cmake", https://wang-yimu.com/a-tutorial-on-shared-memory-inter-process-communication/