Professional Documents
Culture Documents
- Introduction -
Biological inspiration
Animals are able to react adaptively to changes in their
external and internal environment, and they use their nervous
system to perform these behaviours.
Biological inspiration
Dendrites
Axon
Biological inspiration
axon
dendrites
synapses
Dendrites: Input
Cell body: Processor
Synaptic: Link
Axon: Output
Elements of ANNs
Processing Elements (PE)
Processing Element (PE) : PE are artificial neurons
similar to biological neurons.
Each PE receives input, process them and deliver the
output.
Input can be raw input data or output of PE
The output can be the final result.
Or it can be an input to other neurons.
Artificial neurons
Neurons work by processing information. They receive and
provide information in form of spikes.
x1
w1
x2
Inputs
xn-1
xn
z wi xi ; y H ( z )
w2
x3
i 1
..
w3
wn-1
wn
The McCullogh-Pitts model
Output
y
Artificial neurons
The Mc Cullogh-Pitts model:
spikes are interpreted as spike rates;
synaptic strength are translated as synaptic weights;
excitation means positive product between the
incoming spike rate and the corresponding synaptic
weight;
Inputs
Output
Output
y11 2
1
2
2
y1 f ( y , w1 )
y
3
1
1
1
y 2 f ( x2 , w2 ) y 1 y 2 y 2 f ( y 1 , w 2 ) y 2 y 2 y f ( y 2 , w3 )
Out
1
3
1 2
2
1
1
2
y3 y 2 f ( y1 , w 2 )
y3 f ( x3 , w3 )
y3
y1 3
3
4
y14 f ( x4 , w14 )
y11 f ( x1 , w11 )
Neural Networks
highly parallel processing
slow processing units
reliable units
unreliable units
static infrastructure
dynamic infrastructure
prediction
FUNCTION
APPROXIMATION
approximation
tasks.
xm
........
....
Processing
Output
x2
x1
= X1+X2 + .+Xm =y
xm
weights
........
....
wm
...
..
x2
w2
Processing
Output
x1
w1
= X1w1+X2w2 + .+Xmwm
=y
xm
........
....
weights
Processing
Transfer Function
(Activation Function)
Output
wm
...
..
x2
w2
f(vk)
x1
w1
Specifically, by trialanderror
Learning by trialanderror
Continuous process of:
Trial:
Processing an input to produce an output (In terms of
ANN: Compute the output function of a given input)
Evaluate:
Evaluating this output by comparing the actual output with
the expected output.
Adjust:
Adjust the weights.
Learning Paradigms
Supervised learning
Unsupervised learning
Reinforcement learning
Supervised learning
This is what we have seen so far!
A network is fed with a set of training samples (inputs
Unsupervised learning
No desired output is associated with the training
data!
Faster than supervised learning
Used to find out structures within data:
Clustering
Compression
Reinforcement learning
Like supervised learning, but:
Weights adjusting is not directly related to the error
value.
The error value is used to randomly, shuffle weights!
Relatively slow learning due to randomness.
current input;
sends one output signal to many other neurons, possibly
including its input neurons (ANN is recurrent network).
Back-propagation is a type of supervised learning, used
Applications Areas
Function approximation/ Regression
including time series prediction and modelling.
Trained to predict whose values are integer numbers
Data processing
including filtering, clustering blinds source separation and
compression.
(data mining, e-mail Spam filtering)
Applications Areas
Clustering : data set is so complicated that there is no way
Advantages / Disadvantages
Advantages
Adapt to unknown situations
Powerful, it can model complex functions.
Ease of use, learns by example, and very little user
Conclusion
Artificial Neural Networks are an imitation of the biological
Conclusion
Neural networks also contributes to area of research such a
CLASSIFICATION AND
REGRESSION TREES (CART)
Also known as decision trees
Decision Trees:
CLASSIFICATION AND
REGRESSION TREES (CART)
CART aims to use a set of predictor variables to estimate the
means of one or more response variables. A binary tree is
constructed by repeatedly splitting the data set into subsets. Each
individual split is based on a single predictor variable and is
chosen to minimise the variability of the response variables in
each of the resulting subsets.
The tree begins with the full data set and ends with a series of
terminal nodes. Within each terminal node, the means of the
response variables are taken as predictors for future observations.
Closer to ANOVA than regression in that data are divided into a
discrete number of subsets based on categorical predictors and
predictions are determined by subset means.
CLASSIFICATION AND
REGRESSION TREES (CART)
Must define two criteria:
1. A measure of impurity or inhomogeneity.
2. Rule for selecting optimum tree.
Produce a very large tree and then prune it into successively
smaller trees. Skill of each tree is determined by crossvalidation. Divide the full data into subsets, drop one subset,
grow the tree on the remaining data and test it on the omitted
subset.
CLASSIFICATION AND
REGRESSION TREES (CART)
A SUMMARY:
Explain variation of single response variable by one or more
explanatory or predictor variables.
Response variable can be quantitative (regression trees) or categorical
(classification trees).
Predictor variables can be categorical and/or quantitative.
Trees constructed by repeated splitting of data, defined by a simple rule
based on single predictor variable.
CLASSIFICATION AND
REGRESSION TREES (CART)
At each split, data partitioned into two mutually exclusive
groups, each of which is as homogeneous as possible. Splitting
procedure is then applied to each group separately.
Aim is to partition the response into homogeneous groups but to
keep the tree as small and as simple as possible.
Usually create an overlarge tree first, pruned back to desired size
by cross-validation.