ARTIFICIAL NEURAL NETWORKS.ppt on machine learning

SamuelMwangi92 20 views 15 slides May 28, 2024
Slide 1
Slide 1 of 15
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15

About This Presentation

AnN machine learning


Slide Content

ARTIFICIAL NEURAL
NETWORKS
CLASSIFICATION

DEFINITION OF NEURALNETWORK
An artificial neural network (ANN) is a parallel
connection of a set of nodes called neurons.
From the statistical viewpoint, it represents a
function of explanatory variables which is
composed of simple building blocks and which
may be used to provide an approximation of
conditional expectations or, in particular,
probabilities in regression or classification.

ANN WITHANINPUT
LAYER, ONEHIDDENLAYERANDANOUTPUTLAYERI1
I2
I3
I4
Sepal.Length
Sepal.Width
Petal.Length
Petal.Width
H1
H2
O1
O2
O3
setosa
versicolor
virginica
B1 B2

DEFINITIONOFTHEANN
In this study, we only consider:
A feed-forward net with d + 1 input nodes,
One layer of Hhidden nodes,
Coutput nodes ,
An activation function
The input and hidden layer nodes are connected by
weights W
hj for
and
The hidden and output layers are connected by
weights for
and  Hh,...,1  dj,...,0 ch  Hh,...,1  Cc,...,1 )(x

TRAININGTHENETWORK
The connection weights are adjusted through
training.
There exists two training paradigms: Non
supervised and supervised learning.
We discuss and later apply supervised learning.

REQUIREMENTS FORSUPERVISEDTRAINING
A sample ofd input vectors, of size
neach,
An associated output vector,
of size n,
The selection of an initial weight set.
A repetitive method to update the current weights to
optimize the input-output map.
A stopping rule.dXXX ,...,
1 CYYY,...,
1

A REPETITIVEMETHODTOUPDATETHECURRENT
WEIGHTSTOOPTIMIZETHEINPUT-OUTPUTMAP
Error Function
The maximum likelihood method is used to
determine the error function.
The error function is then used to train a given
network.  
2
1
);();,( 


n
i
ii
XZYXYS 

UPDATINGTHEWEIGHTS
This step involves updating the weights until the error
function is minimized.
There are various methods
of minimizing namely:
(1)Backpropagation(BP)
(2)The Quasi-Newton method
(3)Broyden–Fletcher–Goldfarb–Shanno(BFGS)
method
(4)The Simulated Annealing method);,(XYS );,(XYS

NUMBEROFHIDDENLAYERS
A neural net can have more than one hidden layer,
see Looney [1] for more details.
However, it is shown in White [2] that one hidden
layer with sufficient number of neurons is enough
to approximate any function of interest.
In practice, however, a network with more than
one layer may provide a more parsimonious
model for the data.
The number of neurodes, H, in the hidden layer can
be determined as in Looney [2] by a rule of thumb: 1)(log*7.1
2
 nH

NUMBEROFHIDDENLAYERS
Alternatively, one can use the Black Information
Criterion (BIC) as proposed in Swanson and White
[3] to sequentially determine H.

STOPPINGRULES
There are four common stopping rules:
(1)
(2)
(3) a pre-specified lower bound
for the training error
(4) where MAX is the pre-specified
number of iterations.
We note that Rule(4) can be used together with
Rule(1), Rule(2) or Rule(3).0
)()1(


forQQ
rr smallbutforQXYSQXYS
rr
0)
ˆ
;,()
ˆ
;,(
)()1(


 min
)(
)
ˆ
;,( EQXYS
r
 MAXr

NNETINR
They are many packages for ANN. We will use the nnet
package that comes with standard R software.
Run the provided Artificial Neural Net R program. Make
sure you understand each and every step.
Fine tune the program based on:
Number of hidden nodes
Important Variables

EXERCISE
OPTIMIZE THE NETWORK ON THE NUMBER OF
HIDDEN NODES AS INFORMED BY THE AIC
Y –AXIS AIC
X-AXIS NUMBER OF HIDDEN NODES
Hint: You may use the for loop

REFRENCES
(1) Looney, C. G. Pattern recognition using neural
networks: theory and algorithms for engineers and
scientists. Oxford University Press,Newyork, 1997.
(2) White, H. “Some asymptotic results for learning in single
hiddenlayer feedforward network models” [J. Amer.
Statist. Assoc. 84 (1989), no. 408, 1003–1013;
MR1134490 (92e:62119)]. J. Amer. Statist. Assoc.
87, 420 (1992), 1252.

REFERENCES
(3) Swanson, N. R., and White, H. A model-selection
approach to assessing the information in the term
structure using linear models and artificial neural
networks. J. Bus. Econom. Statist. 13, 3 (1995),
265 –275.
Tags