Top Banner
29

Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Jan 19, 2016

Download

Documents

Donald Lawson
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Page 2: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Neural Networks

Presented by M. Abbasi

Course lecturer:Dr.Tohidkhah

Page 3: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Neural networks are adaptive statistical models based on an analogy with the structure of the brain.

Page 4: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

STOP

CoronaryDisease

NeuralNet

Page 5: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Basically, neural networks are built from simple units, sometimes called neurons or cells by analogy with the real thing. These units are linked by a set of weighted connections. Learning is usually accomplished by modification of the connection weights. Each unit codes or corresponds to a feature or a characteristic of a pattern that we want to analyze or that we want to use as a predictor.

Page 6: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Biological Analogy

Synapses

Axon

Dendrites

Synapses++

+--

(weights)

Nodes

Page 7: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Page 8: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Computational Structure of a neuron

. .

.

x1

x2

xN

w1

w2

wN

y

kkk xwy

Page 9: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Multi-Layer Neural Network

Page 10: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

The goal of the network is to learn or to discover some association between input and output patterns, or to analyze, or to find the structure of the input patterns.

Page 11: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

The learning process is achieved through the modification of the connection weights between units. In statistical terms, this is equivalent to interpreting the value of the connections between units as parameters (e.g., like the values of a and b in the regression equation y = a + bx) to be estimated.

Page 12: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Any function whose domain is the real numbers can be used as a transfer function.

Page 13: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

The most popular ones are :

• The linear function• The step function (activation values less than

a given threshold are set to 0 or to −1 , and the other values are set to +1)

• The logistic function [f(x) =1/(1 + exp{-x})]¸ which maps the real numbers into the interval [-1 + 1] and whose derivative, needed for learning, is easily computed {f’(x) = f(x) [1-f(x)]

• The normal or Gaussian function

Page 14: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

The architecture (i.e., the pattern of connectivity) of the network, along withthe transfer functions used by the neurons and the synaptic weights, completely specify the behavior of the network.

Page 15: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Neural networks are adaptive statistical devices. This means that they can change iteratively the values of their parameters (i.e., the synaptic weights) asa function of their performance.

Page 16: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

These changes are made according to learning rules which can be characterized as supervised (when a desired output is knownand used to compute an error signal) or unsupervised (when no such error signal is used).

Page 17: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

The Widrow-Hoff ( gradient descent or Delta rule) is the most widely known supervised learning rule. It uses the difference between the actual input of the cell and the desired output as an error signal for units in the output layer.

Page 18: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Units in the hidden layers cannot compute directly their error signal but estimate it as a function (e.g., a weighted average) of the error of the units in the following layer.This adaptation of the Widrow-Hoff learning rule is known as error backpropagation.

Page 19: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

This adaptation of the Widrow-Hoff learning rule is known

as Error Backpropagation:

2)(2

1odE

•Minimizes the mean squared error using a gradient descent method

Wd

EdWW '

•Error is backpropagated into previous layers one layer at a time.

•Does not guarantee an optimal solution, as it might converge onto a local minimum

•takes a long time to train and requires long amount of training data

Page 20: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Alternative Activation functions

• Radial Basis Functions– Square– Triangle– Gaussian!

• (μ, σ) can be varied at each hidden node to guide training fH(x)

Input 0 Input 1 Input n...

fH(x)fH(x)

fRBF

(x)fRBF

(x)fRBF

(x)

Page 21: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Typical Activation Functions

• F(x) = 1 / (1 + e -k ∑

(wixi) )• Shown for

k = 0.5, 1 and 10

• Using a nonlinear function which approximates a linear threshold allows a network to approximate nonlinear functions

Page 22: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

The Hebbian rule is the most widely known unsupervised learning rule, it is based on work by the Canadian neuropsychologist Donald Hebb, who theorized that neuronal learning (i.e., synaptic change) is a local phenomenon expressiblein terms of the temporal correlation between the activation values of neurons.

Page 23: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Specifically, the synaptic change depends on both presynaptic and postsynaptic activities and states that the change in a synaptic weight is a function of the temporal correlation between the presynaptic and postsynaptic activities.

Page 24: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Specifically, the value of the synaptic weight between two neurons increaseswhenever they are in the same state and decreases when they are in different states.

Page 25: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

What can a Neural Net do?

• Compute a known function• Approximate an unknown

function• Pattern Recognition• Signal Processing• Learn to do any of the above

Page 26: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

·   pattern association ·   pattern classification ·   regularity detection ·   image processing ·   speech analysis ·   optimization problems ·   robot steering ·   processing of inaccurate or

incomplete inputs ·   quality assurance ·   stock market forecasting ·   simulation ·   ...

The areas where neural nets may be useful

Page 27: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

One the most popular architectures in neural networks is the multi-layer perceptron.

Page 28: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Hopfield Net structure

Page 29: Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.

Recap – Neural Networks

• Components – biological plausibility– Neurone / node– Synapse / weight

• Feed forward networks– Unidirectional flow of information– Good at extracting patterns, generalisation

and prediction– Distributed representation of data– Parallel processing of data– Training: Backpropagation– Not exact models, but good at

demonstrating principles

• Recurrent networks– Multidirectional flow of information– Memory / sense of time– Complex temporal dynamics (e.g. CPGs)– Various training methods (Hebbian,

evolution)– Often better biological models than FFNs