GABPN genetic algorithm based back propogation networknew.pptx

ravikumarfulwaria 219 views 25 slides Jul 08, 2024
Slide 1
Slide 1 of 25
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17
Slide 18
18
Slide 19
19
Slide 20
20
Slide 21
21
Slide 22
22
Slide 23
23
Slide 24
24
Slide 25
25

About This Presentation

geentic algorith based back propogation network


Slide Content

Genetic Algorithm Based Back Propagation Network Submitted By: Submitted To: Rohit Poonia Dr.Linni Mathew 231516 ( HOD & Professor, NITTTR) Ravi Kumar Fulwaria 231515 Department of Electrical Engineering, NITTTR Chandigarh

Content Genetic Algorithm Back Propagation network Genetic Algorithm based Back Propagation Network Coding Weight Extraction Fitness Factor Reproduction Convergence Conclusion References

Genetic Algorithm The Genetic Algorithm is a method for solving both constrained and unconstrained optimization problems that is based on natural selection, the process that drives biological evolution. The Genetic Algorithm repeatedly modifies a population of individual solutions. At each step, the genetic algorithm selects individuals from the current population to be parents and uses them to produce the children for the next generation. Over successive generations, the population "evolves" toward an optimal solution.

Flowchart of Genetic Algorithm

Back Propagation Algorithm in Neural Network In an   artificial neural network, the values of weights and biases are randomly initialized. Due to random initialization, the   neural network probably has errors in giving the correct output. We need to reduce error values as much as possible. So, to reduce these error values, we need a mechanism that can compare the desired output of the neural network with the network’s output that consists of errors and adjust its weights and biases such that it gets closer to the desired output after each iteration. For this, we train the network such that it back propagates and updates the weights and biases. This is the concept of the back propagation algorithm.

Flowchart of Back Propagation Algorithm

Genetic algorithm (GA) based Back Propagation networks (BPN) Conventionally, a BPN determines its weights based on a gradient search technique and therefore the risk of encountering the local minimum problem. GA on the other hand, though not guaranteed to find global optimum solution to the problems but have been found to be good at finding “acceptably good” solutions to problems at the quicker rate or use less number of iterations. The idea to hybridize the two approaches, namely Genetic Algorithm and Backpropogation networks was given by Whitley , Bogart and Hanson in year 1990. They use GA to guide BPN in finding the optimal weights and to enhance the speed of training.

Flowchart of Genetic Algorithm ) based Back Propogation networks

Genetic Algorithm based Back Propagation Network has 5 parts Coding Weight Extraction Fitness Factor Reproduction Convergence

1.Coding Chromosomes: The parameters which represent a potential solution to the problems, (genes) are joined together to form a string of values referred to as a chromosomes. Assume a BPN whose network configuration is l-m-n ( l input neurons, m hidden neurons and n output neurons). The number of weights that are to be determined are ( l + n )* m . Let d = The number of digits in the weight Then String Length ,L = ( l + n )* m*d is randomly generated The string S represents the weight matrices of the input-hidden and hidden-output layers in a linear form. An initial population of “ p” chromosomes is randomly generated where “ p” is referred as the “Population Size”

Example Of Coding: Consider a BPN network with configuration 3-2-1: l=3 input neurons, m=2 hidden neurons and n=1 output neurons The number of weights that are to be determined are ( l + n )* m = (3+1)*2= 8 Let d = 5 The number of digits in the weight Then String Length ,L = ( l + n )* m*d = 8*5=40 is randomly generated

2. Weight Extraction To determine the fitness values for each of the chromosomes, we extract weights from each of the chromosomes. Let x 1 ,x 2 ,……. x d ,……,.. x L represents a chromosome and x kd+ 1, x kd+2 ,.. …….x (k+1)d represent the k th gene (k≥0) in the chromosome. The actual weight w k is given by + x kd+2 10 d-2 + x kd+3 10 d-3 +……. x (k+1)d , if 5≤x kd+1 ≤ 9 w k 10 d-2 - x kd+2 10 d-2 + x kd+3 10 d-3 +……. x (k+1)d , if 0≤x kd+1 ≤ 5 10 d-2 Let chromosome of length 40 : 84321 46234 78901 32104 42689 63421 46421 87640 For Gene 0: 84321 we have k=0, d=5 and x kd+1 = x 1 which is = ‘8’ such that 5≤x 1 =8 ≤ 9 hence the weight extracted is w = + 4X10 3 + 3X10 2 + 2X10 +1 = 4.321 10 3

Let chromosome of length 40 : 84321 46234 78901 32104 42689 63421 46421 87640 For Gene 1: 46234 we have, k=1, d=5 and x kd+1 = x 6 which is = ‘4’ such that 0≤x 6 =4 ≤ 5 hence the weight extracted is w 1 = - 6X10 3 + 2X10 2 + 3X10 +4 = -6.234 10 3 Similarly, Gene 2: 78901 yields w 2 = +8.901 Gene 3: 32104 yields w 3 = -2.104 Gene 4: 42689 yields w 4 = -2.689 Gene 5: 63421 yields w 5 = +3.421 Gene 6: 46421 yields w 6 = -6.421 Gene 7: 87640 yields w 7 = +7.640

3. Fitness Function Fitness Function (also known as the Evaluation Function) evaluates how close a given solution is to the optimum solution of the desired problem. It determines how fit a solution is. The fitness function must be calculated for each problem to be solved. The root mean squared error  E i  of an individual program  i  is evaluated by the equation: where P( ij ) is the value predicted by the individual program  i  for fitness case j (out of n fitness cases); and  T j  is the target value for fitness case j. For a perfect fit, P ( ij )  =  T j  and  E i  = 0. So, the RMSE index ranges from 0 to infinity, with 0 corresponding to the ideal. Thus, for evaluating the fitness  F i  for the chromosome  i , the following equation is used F i = 1 E i  

4. Reproduction In this phase the mating pool is first formed before the paranet chromosomes reproduce to deliver offspring with better fitness. For the given problem the mating pool is first formed by excluding that chromosome C i with the least fitness F min and replacing it with a duplicate copy of the chromosome C k reporting the highest fitness F max . Now the best fit individuals have multiple copies while the worst fit individuals die off. From the mating pool the parents are selected in pairs at random The chromosomes of the respective pairs are recombined using the “Two point cross over” operator of a standard GA

Two-Point Crossover This is a specific case of a N-point Crossover technique. Two random points are chosen on the individual chromosomes (strings) and the genetic material is exchanged at these points.   The offspring which now form the current population again have their fitness calculated as before by fitness function

5.Convergence For any problem, if the GA is correctly implemented, the population evolves over successive generations with the fitness value increasing towards the global optimum value. Convergence is the progression towards increasing uniformity. A population is said to have converged when 95% of the individuals constituting the population share the same fitness value. The population P 1 now undergoes the process of selection, reproduction and cross over. The fitness values for the chromosomes in P 1 are computed, the best individuals replicated and reproduction carried out using two-point cross over operator to form the next generation P 2 of chromosomes. The process of generation proceeds until at one stage 95% of the chromosomes in the population P i converge to the same fitness value. At that stage the weights extracted from the population Pi are the final weights to be used by the Back Propagation Network

Example

Limitations and Challenges Vanishing and Exploding Gradients: In very deep networks, gradients can become very small (vanish) or very large (explode), making training difficult. Techniques like LSTM in RNNs or ResNet in CNNs are used to mitigate this. Requirement of Differentiable Activation Functions: The activation functions used in the network must be differentiable for backpropagation to work.

Limitations and Challenges Vanishing and Exploding Gradients: In very deep networks, gradients can become very small (vanish) or very large (explode), making training difficult. Techniques like LSTM in RNNs or ResNet in CNNs are used to mitigate this. Requirement of Differentiable Activation Functions: The activation functions used in the network must be differentiable for backpropagation to work.

Limitations and Challenges Vanishing and Exploding Gradients: In very deep networks, gradients can become very small (vanish) or very large (explode), making training difficult. Techniques like LSTM in RNNs or ResNet in CNNs are used to mitigate this. Requirement of Differentiable Activation Functions: The activation functions used in the network must be differentiable for backpropagation to work.

Limitations and Challenges Vanishing and Exploding Gradients: In very deep networks, gradients can become very small (vanish) or very large (explode), making training difficult. Techniques like LSTM in RNNs or ResNet in CNNs are used to mitigate this. Requirement of Differentiable Activation Functions: The activation functions used in the network must be differentiable for backpropagation to work.

Limitations and Challenges Vanishing and Exploding Gradients: In very deep networks, gradients can become very small (vanish) or very large (explode), making training difficult. Techniques like LSTM in RNNs or ResNet in CNNs are used to mitigate this. Requirement of Differentiable Activation Functions: The activation functions used in the network must be differentiable for backpropagation to work.

Limitations and Challenges Vanishing and Exploding Gradients: In very deep networks, gradients can become very small (vanish) or very large (explode), making training difficult. Techniques like LSTM in RNNs or ResNet in CNNs are used to mitigate this. Requirement of Differentiable Activation Functions: The activation functions used in the network must be differentiable for backpropagation to work.

Thank You!!
Tags