Definition of an ANN (Artificial Neural Networks)
An ANN is completely specified once we define its:
1. Architecture: An ANN is a directed or un-directed graph where each node (or vertices) is called a neuron and each arch (or edge) is called a synaptic connection, it has three subset: the input units, the output units and the hidden units. β For each node we can define an activation function β While for each arch we can define its weight. An architecture is completely defined with: Vertices, Edges, Input Units, Output Units, Hidden Units, Weights and Activation Functions.
1.1. Typical Activation Functions: β Step function or TLU (Threshold Logic Units): β Linear Function: β Sigmoid function: β Hyperbolic tangent sigmoid: β Gaussian: β ReLU (Rectifier Linear Units): β Leaky ReLU:
2. Dynamics: How the signal propagates. The dynamics of an ANN represent how the input data goes from start to end (how it βpropagatesβ), we need to define how the weights interact with the data after they are processed by the activation function. ~Ex.: Letβs take input data these data will not have any transformation in the input (even this part can be changed), then they will pass through a first hidden layer where for each node the activation function will be something like:
Where: : old layers (from to ) (in this case the input) : node belonging to the new layer (in this case the hidden layer) : sigmoid function (our chosen activation function)
~Ex.:

This process is then repeated until the signal reaches the output layer.
Also the dynamics also define the clock (time trigger) of the ANN, which depends on its family of networks.
NOTE: The ANN topology specifies the hardware architecture, the value of the weights itβs the software, while the ANN dynamics (the living machine) represents the running process.
3. Learning: The ANN learns from the examples contained in the training set which is a continuous-valued data sample drawn from an underlying multivariate pdf. Main learning setups: β Supervised Learning: β Unsupervised Learning: β Semi-Supervised Learning: training data are supervised, they are a set of input and output , while data are unsupervised, we know the input but not its output β Reinforcement Learning: Where a reinforcement signal either a penalty or a reward is given every now and then.
3.1. Generalization: Learning is a process of progressive modification of the connection weights, aimed at inferring the (universal) law underlying the data. Far from being a mere memorization of the data, the laws learned this way are expected to generalize new data, previously unseen, this is called generalization capability.
Original Files:


