Top Banner
Complex Systems 7 (1993) 445-467 U sing the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training N achum Shamir* David Sa ad! Emanuel Marom Faculty of Engineering, Tel Av iv University, Ramat Aviv 69978, Israel Abstract. We propose a new hybrid genet ic back propagation tr ain- ing algorithm based on a unique f unct ional matching recombination method. The met hod is used to evolve pop ulat ions of neur al networks and provides versatility in network architect ure and activation func- tions. Net reorganization and reconstruction is carried out prior to genet ic recomb ina tion usin g a funct ional behavior correla tion measur e to compare the functional role of the various neurons. Comparison is done by correlating the internal representations gen erat ed for a given tr aining set. Net structure is dynamically changed during the evolu- tio nary process, expanded by reorganization and reconstruction and tr imm ed by pruning unnecessary neurons. The ability to change net structure througho ut generations allows the net population to fit itself to the requireme nts of dy namic adapt ation, performance, and size con- siderations in the selection process, thus gene rat ing smaller and more efficient nets t hat are likely to have higher generalization capabilities. A func t ional behavior correlation meas ure is used exte nsively to ex- plore and compare nets and neurons, and its ability is demonstrated by investigating the results of genet ic recombinat ion. The vitality of nets organized via the functional behavior correlation measure prior to genetic recombination is demonst rated by stat ist ical results of com- puter simulations. The performance of the proposed method and its generalization ca pabilities are demo nstrated using Parity, Symmetry and handwritten digit recognition training tas ks. "Current address: Department of Electrical Engineering, Technion-I srael Institute of Technology, Technion City, Haifa 32000, Israel. tCurrent address: Department of Physics, University of Edinburgh, J. C. Maxwell Building, Mayfield Road, Edinburgh EH9 3JZ, UK.
23

Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Jun 03, 2020

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Complex Systems 7 (1993) 445-467

U sing the Functional Behavior of Neurons forGenetic Recombination in Neural Nets Training

N achum Shamir*David Saad!

Emanuel Maro mFac ulty of Engineering, Tel Aviv University,

Ram at Aviv 69978, Israel

Abstrac t . We propose a new hyb rid genetic back propagation train­ing algorithm based on a unique functional matching recombinationmethod. T he method is used to evolve pop ulations of neur al networksand provides versatility in network architecture and activation func­tions. Net reorganization and reconstruction is carried out prior togenet ic recomb ina tion usin g a funct ional behavior correlation measureto compare the functional role of the var ious neurons. Compar ison isdon e by correlating the intern al representations generated for a giventraining set. Net st ruc ture is dynamically changed during the evolu­tionary process, expanded by reorganization and reconstruct ion andtrimmed by pruning unnecessary neur ons. The ab ility to change netstructure throughout generations allows t he net population to fit itselfto the requirements of dynamic adaptation , performance, and size con­siderations in the selection process , t hus generating smaller and mor eefficient nets t hat are likely to have higher generalization cap ab iliti es.A func t ional behavior corr elation meas ure is used exte nsively to ex­plore and compare nets and neurons, and its ability is demonst rat edby investi gat ing the results of genetic recombination. The vit ality ofnets organized via the functional behavior correlat ion measure pr iorto genet ic recombinati on is demonstrated by statist ical results of com­

puter simulat ions. The performance of t he proposed method and itsgeneralization capabilit ies are demonstrated using Parity, Symmetryand handwrit ten digit recognit ion training tasks.

"Current address: Department of Electrical Engineering, Technion-I srael Institute ofTechnology, Technion City, Haifa 32000, Israel.

tCurrent address: Department of Physics, University of Edinburgh, J. C. MaxwellBuilding, Mayfield Road, Edinburgh EH9 3JZ, UK.

Page 2: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

446 Nachum Shamir, David Saad, and Em anuel Marom

1. Introduction

Const ruct ive and dest ruct ive tra ining algorithms for neural nets have signif­icant importance because of their ability to construct minimal nets that areeconomical in te rms of hardware and software and powerful in ter ms of gen­era lization capabilit ies. This work presents a novel approach for combininggenetic evolut ion, back propagat ion t ra ining, and various pruning methodsto provide a powerful t ra ining algorithm capable of dynamically modifyingnet struct ure , functional composit ion, and weights , while adapting towardminimal net st ructure .

The basic concepts of construction algorithms are demonstr ated by theTiling [7] and Upstart [8J algorithms that create a feedforward network ofbinary neurons and a single output neuron. The more advanced CascadeCorre lation [9, 10] algorithm uses continuous neurons with no limitat ion onthe numb er of output neurons. In these algorithms, a combined tra iningalgorithm and construction operator provide the capability of building the netgradually. Neurons and layers are added to the net as needed and convergenceis guarant eed regardless of t he initi al net st ructure .

Such "forward progressing" algorithms suffer from significant drawbacksin their ability to produce minim al nets. It has been suggested [1, 4, 5, 6] thatthe most efficient and promising way to pro duce a minimal net is to t ra in asufficient ly large net that is pruned both during and after training. A "for­ward progressing" algorithm is likely to produce nets that have more neuronsand layers than act ua lly needed; bu t due to the nature of t he constructionoperator , it is difficult to remove redundant uni ts.

Extensive use of genet ic algorithms for neur al net tr aining and const ruc­tion has been shown to improve exist ing t ra ining techniques and overcomecertain limit at ions such as local minima traps, network paralysis, and others .Genetic algorit hms are based on various aspec ts of natural select ion and ge­netics. Taking an opt imization problem and encoding its solut ion proposalsinto a populat ion of artificial "chromosomes," one makes use of select ion andreproduct ion operators similar to natural ones, and evolves a populat ion ofsolut ion proposals. Using effective natural select ion criteria and efficient re­produ ct ion methods, the popul at ion is enhanced with each generation. T heresult is a powerful stochas t ic optimization t echnique.

T he applicat ion of genet ic algorithms for neur al net t raining is done in avariety of ways. In common applications, demonstrated successfully by theGEN ITOR and GENITORII algorithms [12, 13, 14], network weights areencoded into art ificial "chromosomes" represented by bin ary strings. Con­vent ional tr aining techniques (rest ricted to a predefined network st ructure)are then applied. A more globa l approach that encodes net connectiv ity wasdemonst rat ed by Whi tley et al.[14], in which an evolut ionary pro cess wasused to create the interconnections map , and conventional training methodswere used to train the proposed nets. These training algorithms are confinedto a predefined network architecture where the number of layers and neuronsdo not change during training.

Page 3: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Genetic R ecombin ation 447

An example of a genetic algorit hm th at searches for an appropriate net­work architecture was presente d by Kitano [15]. Recognizing t he "scalability"probl em of methods based on direct encoding of the interconnections map ,Kit ano proposed the encoding of a graph generation gra mmar that encodesthe network const ruc t ion rules into th e artificial chromosomes. The encodedgramm ar was significant ly shorter tha n direct encoding methods, and pro­vided the means for dealin g with dyn amically changing configurati ons. Itis important to note that every offspring net crea ted by this algorit hm wastrained from a random initial state . Besides connect ivity pat terns, no weightvalues were tr ansferr ed to the offspring net s and an expensive (in terms ofcomputational complexity) back propagation t raining was used. In a pro­cedure similar to the introduction of priors a par ticular configuration waschosen from a cert ain class of possible solutions. The exac t solut ion, that is,the explicit weight matrix, is defined by a complementary t raining process(such as gradient descent ) tha t refines the solut ion within th e given class ofsolut ions.

We propose a new functional mat ching recombinati on method for usewith genetic back propagation hybrids, where matching is done by comparingthe functional role of neuron pairs correlated by their corres ponding inter­nal repr esentations. The represent ation of neuron functionalit y by a vectorof intern al representations for the ent ire training set is called the junctionalbehavior o] neurons [1, 2] and is also used for observing the results of genet icrecombination. The proposed recombination method t akes int o account thefact that neurons performing equivalent tas ks may be located at differentpositions in the hidd en layers and therefore encoded at different locati ons inthe genetic "chromosomes" (this is often called the problem oj hidden neu­ron location permutat ion [20, 21]). By rearranging and reconstructing theparent nets prior to geneti c recombinati on , the recombination efficiency issignificantly enhanced and vital intern al st ruct ures are preserved. In addi­tion , since parent nets are reconstructed prior to genet ic recombination, thepopulation may include a variety of network st ruc tures.

The main cont ribut ion of the proposed method is it s ability t o handlehet erogeneous populations of nets having different sizes and st ructures . It sability to t ra nsfer network infrastructures and their corresponding weightvalues from parents to the offspring nets is also recognized. This enables asmoot her evolution of the population between different classes of solut ions(mainly configurat ions and act ivation functions) , thereby creating offspringnets with enhanced initial performance that require fewer retraining cycles.By adding a pruning phase, the hybrid algorithm adapt ively cha nges thest ruct ure of the nets in the population. These changes are cont rolled by abalance between expa nsion and pruning pro cesses.

2. Genetic algorithms

Genetic algorit hms are st ochas t ic optimization methods that imitat e naturalpro cesses by applying "evolut ion" to a population of solut ions proposed for

Page 4: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

448 Nachum Shamir, David Saad, and Emanuel Marom

a given optimi zati on t ask. Each proposal in th e population is given a fitnessvalue representing its performance when applied to a specific task. Duringevolut ion, pairs of individu al solut ions are chosen from the population (atrandom), and together they produce an offspring representing a new solut ionproposal. The offspring is given a fitness value with which it competes fora place in the population. Similar to biological natural select ion, less fitindividu als generally do not survive and are removed from the population.The fitn ess of t he ent ire populati on is enhanced with each generation until themost fit individuals reach the global optimum. Several variat ions of geneticopera to rs have been suggested and to clar ify the typ es of operators used int his work, we provide the following brief review.

Each individu al "chromosome" is represent ed by a st ring of encoded fea­tures and parameters. Then the reproduction stage mentioned above is car­ried out in the following mann er.

• Two parents are chosen at random from the population , using a non­uniform prob abili ty dist ribu tion. Higher prob ability is given to more fitindividu als, thus adding selective pressure [11, 13] to the evolut ionaryprocess.

• A new offspring is crea ted by combining attributes (selected at ran domwith equal probability) from each of the two parents.

• Mut ations are applied to the offspring by making small random changesto the encoded features.

• According to the principle of natural select ion, the new offspring mustcompete for a place in the population. A fitness value is given to t henew offspr ing , which is then compared to the least fit individual in t hepopulation, result ing in th e surv ival of the more fit individual and theeliminat ion of the ot her .

The parameters of thi s art ificial evolut ionary pro cess must be selectedcarefully. The nonuniform distribution underlying t he select ion of parentsmust be carefully chosen to avoid exaggerated preference of highly fit indi­vidu als which may cause a premature popul ati on convergence and loss ofdiversity. On the other hand, if more fit individu als are not gra nted a statis­tical preference, the convergence rat e is slowed and pro cessing t ime is wasted.The rate of mutation must also be carefully considered. High mutation ratesmay slow the evolutionary pro cess by creat ing a mass of low-fit ness individ­uals that can dest roy the evolution pro cess by transforming it into a randomsearch. An ext remely low mutation rat e, on the other hand , may result inthe loss of population diversity and prevent convergence in the vicinity of th eglobal optimum.

When genet ic algorithms are used for evolving populations of neural net­works, ma jor encoding difficulties are encountered. Neurons that performequivalent t asks wit hin different networks are located at different positions in

Page 5: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neuron s for Genetic Re combination 449

th e hidd en layers and therefore encoded at different locations in the "chromo­somes" of a selected pair of parent nets. Recombining the two parent "chro­mosomes," which are arbit ra rily ordered, may produce meaningless randomresults as well as untrainable offspring nets. This obstacle is often called theproblem of hidden neuron location permutation and is known to have a disas­trous effect on genet ic algorithm training proc esses [20, 21]. The dam age isusually caused by the destruction of vital infrastructures of both parent netsduring recombination, since weight values belonging to different infrastruc­tures are mixed. To minimize t he damage during recombin ation, one mustevaluate the infra structure functional similarity in both parent nets and en­code the connectivity pa tterns and weight values of similar infrastructures atsimilar locations in the parent "chromosomes." Since neuron act ivation func­tions are nonlin ear , it is impossible to evaluate the similarity of neurons andnetwork infrastructures by comparing network connectivi ty or weight val­ues. In th e next section, we present a measur e for evaluat ing infrastructurefunctional similarit ies.

3. Functional behavior of neurons

The "funct iona l behavior" of a neuron describes how that neuron and itscorresponding subnet respond when variou s inpu t vectors are presented tothe net. The net is fully forward connected and has no feedback connections;weight values are real and neuron act ivat ion functions are cont inuous. Eachsubnet is a part of the net start ing at an inpu t layer and ending at a singlehidden or output neuron th at is th e output neuron of th at subnet . The sub­net contains all relevant neurons in previous layers and all interconnectionsleading to those neurons. Every subnet repr esent s a real function f: ~n ---+ ~

on the set of input vectors where n is the number of input neurons. (Not ethat a binary representation is a special case of the general continuous rep­resent ation.) This function is the subnet response function. The output ofthe sub net ending at neuron i is represented by

(1)

where VI , . . . , Vn are th e components of the input vector and Si is the outputvalue of neuron i. The neuron 's functional behavior is defined as the responseof the corresponding subnet to the set of input vectors

(2)

where p is the numb er of input vectors in the dat a set and sj represents theoutput value of neuron i when input vector j is present ed to th e net .

To compare different neurons and functional behaviors, the latter is nor­malized with respect to its overall norm E i = L~=I (Sj)2 , that is,

_ (Si Si ). I p

B ' = -lEi'" '' -lEi . (3)

Page 6: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

450 Nachum Shamir, David Seed, and Emanuel Marom

This norm alized representation simplifies both graph ic and numerical func­t ional comparisons. The degree of matching between a pair of neurons i 1

and i 2 is given by the correlation of their corresponding normalized func­t ional behaviors:

(4)

Thi s normalized matching factor lies in the interval [- 1,1]. Its magnituderepresents the funct ional similarity of the corresponding subnets, where thenegative sign indicates an opposite response. For linearly dependent func­tional behavior vectors, the matching fact or is eit her 1 or -1:

where 0: denotes the linear dependence.

4. G enetic r ecombination opera tor

This secti on proposes a new genet ic recombination operator based on thefunctional behavior correlation measure. The main obj ecti ves of the prop osedoperator are to increase neur al net recombination efficiency by rearrangingthe intern al st ruct ure of parent nets, providing the capability to recombineparent nets of different sizes and thus making the handling of heterogeneousnet popul ations possible. The two parent nets undergo internal rearrange­ment and reconstruction, creating a pair of target nets that have identi calarchitecture and well-matched intern al order.

The mutual reordering process and reconstruction of the parent nets isperformed simultaneously on the two nets, layer by layer.

1. Input layer : The functional role of every inpu t neuron is uniquelydefined by its location in t he input layer, and all inpu t neurons arecopied to the same locat ions in the target nets.

2. Hidden layer: There is no dependency between t he functional role ofhidden neurons and th eir locat ions in the hidden layers. Therefore onemust identi fy related neurons in both nets and place them in the samelocat ion in the hidden layers of the corresponding target nets. Neuronsare copied to target nets togeth er wit h their entire set of int ernal pa­ramete rs, input connect ions, and weight values.Hidden layer process ing is done in th e following mann er:

• All neurons from th e current ly pro cessed hidden layer of the firstnet are copied to the corresponding target net .

• For each neuron in the first net , the neuron wit h t he highest func­t ional similarity is identified in th e second net and copied to the

Page 7: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Genetic Recombination 451

equivalent location in th e second target net , parallel to th e orig­inal neuron in th e first net . The functional similarity of neuronsis evaluated by equation (4), th e functional behavior correlationmeasure.

• If the matching factor (4) is negat ive, invert the activation func­tion and th e outgoing connections of th e second neuron. Whenthe act ivat ion function is ant isymmetric, one simply inverts theincoming connections instead of the act ivat ion function.

• When all neurons from the first net have been pro cessed , repeatthe process for t he second net by identifying and copying neuronswith th e highest funct ional similarity from the first net .

• During th e reord ering pro cess and reconstruction, neurons are du­plicat ed and may therefore appear more than once in the hiddenlayers of th e target nets. The functional performance of the targetnet s is restored by comp ensating the outgoing connect ions of th eduplicat ed neurons. The compensat ion is carried out by dividingth e amplitudes of th e corresponding output connections by thenumb er of duplications.

The number of neurons in the hidden layer crea ted in the target netsis t he sum of hidden layer sizes of th e two original nets.

3. Out put layer: The functional role of each output neuron is uniquelydefined by its location in the output layer. All output neurons arecopied tog ether wit h their ent ire set of int ern al parameters, input con­nections, and weight values, to the same locations in th e t arget nets.

After t he two parent nets are reordered and reconstructed , and the targetnets have been crea ted, an offspring net is creat ed. Each building block ofthe recombinat ion pro cess consist s of a neuron, its internal par ameters (forexample, an act ivat ion function) , a set of input connec tions , and the corre­sponding weight values. It th erefore represents a specific function performedby that neuron (on the spa ce of net inputs as well as outp ut of neurons inprevious layers). The function is kept int act throughout recombination andtransferr ed to th e offspring net. The offspring net is created by parsing thetwo reconstructed nets simultaneously, select ing at random which of themwill supply the next building block , which is then copied to the same lo­cation in the offspring net . Mutat ions are applied by adding random noiseto the offspring weights. The offspring net is retrained and submitted topruning [1, 4, 5, 6] and ret raining cycles.

5. The hybrid training system

We will now describ e an ent ire training algorit hm utilizing genetic algorit hms,back propagation , pruning, and the recombination operator defined in sec­tion 4. Training is divided int o two st ages:

Page 8: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

452 Nachum Shamir, David Saad, and Em anuel Marom

1. initial population generat ion; and

2. genet ic population evolut ion.

Initial population generation is done by training, pruning, and ret rainingnets seeded wit h rand om initial weights. The meth od used for training isback prop agation with moment um (ex = 0.4) and weight decay (E: = 0.005).Back propagat ion t raining continues unt il all out put polarit ies are corrector the epoch limit of 100 epochs is reached, whichever comes first . Networkpruning is done using three different methods:

1. neuron merging using functional behavior corre lation to find and mergematchable pairs of neurons [1];

2. neuron pr uning by removing irrelevant neurons (see, for example, [5]);and

3. interconnect ions pruning by measuring the interconnection relevanceand removing irrelevant interconnections (see, for example, [6]) .

Functional matching is evaluat ed by (4) for all pairs of neurons, and thosepairs having matching magnitudes great er th an or equa l to 0.75 are marked asmerging candidates. This parameter has been determined experimentally [1].

The relevan ce of each neuron k is evaluated by calcu latin g the influenceof its removal on net output

q P

Rk = L L (0 ; - 0 (k); )2;= 1 j=l

(6)

(7)

where OJrepresents t he net outp ut bit i when input vecto r j is fed to the net .

The term O(k); is the net output generated for the same data when neuronk is removed. T he constan t p is the number of vectors in the data set , andq is the number of output neurons. In our simulat ions we used t he relevancemeasure defined at equation (6), but it should be noted that the amount ofcomputat ion can be reduced by using approximations such as those describ edin [5, 6, 17J.

The relative relevance of neuron k is defined by:

RkPk = 1 N

N 2:;=1R;

where N is the total numb er of neurons. The relat ive relevance of intercon­nections is calculated similarly. In all experiments, we used a relat ive neuronrelevance pruning threshold of 0.3 and a relative int erconnect ions relevancepruning threshold of 0.2. The three pruning methods are used together:the merging and pruning candidat es are first marked, and then the nets arepruned and retrained.

The fitness of each individu al net is determined by counting the numb erof correct outputs generated for t he set of t ra ining or testin g vectors. When

Page 9: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Genetic Recombination 453

the fitness of two nets is equal, priority is given to the smaller net . Thenumb er of initi ally tr ained nets must be larger than t he target populationsize. This ensures populations of nets with enhanced initi al perform anceexceeding that of random net selection.

Once the initi al population is created, th e evolut ionary pro cess may be­gin. For each evolutionary step , two individu al nets are chosen at randomfrom th e population such that fitt er individu als have a greater chance of be­ing selected (thi s is called selective pressure [11, 13]). The selected pair ofindividuals recombines and crea tes a new offspring as describ ed in secti on 4.The new offspr ing is trained , pruned, and retrained. Its fitness is calcul at edand compared to the most similar net in the population, which is replacedif the offspring is found to be more fit. Such a method of select ion [2] showsbetter perform ance and preserves th e diversity of an evolving pop ulat ion ofneural networks.

6. Experimental results

The performance of the proposed hybrid training system is illustrat ed by fourexamples. In the first example, the influence of functional reorganization ongenet ic recombination efficiency is examined by comparing the recombin ationperformance with and without reorganization, using the Parity-6 dat a set .In the second example, the genera lization enhancement prop erties of theproposed syste m are demonstrat ed using the Symm etry-16 dat a set , whoseperform ance is comp ared to that of a computationally equivalent non-genet ictraining program. The third is a unique experiment involving heterogeneouspopulations of nets, each implementin g a variety of decision functions. Inthe last experiment, nets are t ra ined to identify handwritten digit s.

6.1 The influence of functional reorganization on geneticrecombination efficiency

This experiment was designed to isolate and investigat e the effects of hidd enneuron funct ional reorganization on the efficiency of genetic recombin ation.Two typ es of recombination methods were compared. In the first , hidd enneuron s were not reorganized prior to recombin ation; in th e second, hidd enneurons were reorganized as describ ed in section 4. All other recombin ationand retraining parameters, including net growt h factors, were identi cal in thetwo experiments.

The population included ten nets. All nets were fully forward connectedand had an initial st ructure of 6:6:6:1. The nets were exhaustively t rainedby the Pari ty-6 data set , subjected to pruning (by the three methods dis­cussed in sect ion 5), and retrained. The initi al population evolved for 2000genera t ions. Population evolut ion was performed according to the guide­lines describ ed in sect ion 5, using all three pruning methods and functionallymatched recombin ation. It is import ant to not e that all nets in the pop­ulatio n had perfect performance throughout t he generations, since all nets

Page 10: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

454 Nachum Shamir, David Saad, and Emanuel Marom

in the init ial pop ulation had perfect performance and t hese were replacedby offspring nets that had equivalent or better scores. Both recombin ati onexperiments were performed using the same populat ion of nets. For the firstpair of tes ts we used the nets of the initial population, for the second pair weused the nets of the populat ion after 1000 genera t ions of evolut ion, and forthe last pair of tests we used the nets of the population after 2000 genera t ionswere performed.

Recombinat ion efficiency was measured by the number of retrainingepochs needed to achieve perfect performance of an offspring. An adequaterecombin ation meth od was one that efficiently t ransferred vital infrastruc­tures from parent to offspring nets, giving high initial perform ance to theoffspring and reducing th e amount of retraining. Note that dur ing recombi­nat ion tests offspring were not subjected to pruning and did not participatein the evolut ion of the pop ulation.

For each histogram in Figures l (a) and l (b), 900 recombin at ions weremade, each tim e selecting a pair of nets from the pop ulati on (with uniformprob abi lity) and ret raining th e newly created offspr ing. The results are dis­played as a histogram of the requir ed number of ret raining epochs neededto achieve perfect training , where retraining was limited to 100 epochs. Thepeaks shown in the histograms at 100 epochs represent the amount of untrain­able offspring created by the corresponding recombina tion meth od , while thepeaks at zero epochs represent offspring created with perfect init ial perfor­man ce. The experiments were performed three t imes: for non-evolved, for1000 generation-old, and for 2000 genera t ion-old populat ions, exploring theeffect of population aging upon th e performance of the correspo nding recom­bination met hod. The results of recombination witho ut functional reorga­nization are displayed in Figure l (a), while the results for the funct ionallymatched recombinat ion meth od are displayed in Figure l (b). For t he initi alnon-evolved pop ulat ion, the two experiments presented in Figure l (a) andFigure l (b) show a minor advantage for the functio nally matched recombi­nat ion method: the numb er of untrainable nets is 137 (out of 900), while 153of t he nets were untrainable for th e recombin ation without functional reor­ganizat ion. The convergence properties of the functionally matched recom­bination method are clearly bet ter for the 1000 and the 2000 generation-oldpopulations.

As popu lat ions grow older , superior infrastructures gradually dominat eth e "chromosomes." Recombin ation without functional reorganizat ion failsto preserve and transfer th ese superior st ructures to the created offspring, asis evident by th e fact that th e histograms have similar shapes (Figure 1(a) ).We find t hat vital infrastruct ures are destroyed during recombination , re­training success is reduced, and the numb er of untrain able offspring is in­creased. The functionally mat ched recombinat ion method, on th e ot her hand,preserves th ose infrastructures as the population evolves. The numb er of netsthat require less retraining to achieve the desired performan ce is increased ,as is the numb er of nets with perfect initial performan ce (represented by th epeaks at zero epochs in Figure 1(b)) .

Page 11: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Geneti c Recombination 455

Retraining epochs histogram for non-evolved population .

5025 75 100Retrainin g epochs

Retraining epochs histogram for 1000 generation-old population.

16

12J!l"c:: 80'+-<0

'* 40

00

75 100Retrain ing epochs

75 100Retrainin g epochs

50

50

25

25

Retra ining epochs histogram for 2000 generation-old population.

'""c::13'+-<

0

'* 65

00

28

'"21

"c::14'+-<

0

'* 70

00

Figure 1: Recombination performance benchmark. Recombinationperformance was t este d for t he initial, 1000 generati on-old , and 2000generation-old populations of ten nets, perfect ly t rained by the Parity­6 dat a set. Each of the t hree tes ts included 900 independent recombi­nations and retrainings. The histo grams show the dist ribution of thenumber of ret raining epochs required to restore perfect performan ceof the genera ted offspring. The peak at 100 epochs repre sent s thosenets that could not be retrained , while the peak at zero epochs (if any)repr esent s nets that were produced wit h perfect initial performance.(a) Performance benchmark for recombinat ion without functional re­organization.

6.2 Generalization enhancement

Generalization cap abi lity is the most important prop erty of neur al networks,and efforts are being made to improve it . Learning from a set of examples(called the training vectors) , nets are capable of responding corre ct ly to newinputs that were not present ed previously during training. However, not all

Page 12: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

456 Nachum Shamir, David Seed, and Emanuel Marom

Retraining epochs histogram for non-evolved population.

5025 75 100Retraining epochs

Retraining epochs histogram for 1000 generation-old population.

14

10~t:: 704-<0

'**' 35

00

15lJg'0 10

'**'

75 100Retraining epochs

5025

o -l.lli1illilltJlliIJ:l:I:H:nIT~=dbill!D..,""""",,,==.~......J:b-=>-_--.--_ _ ~_~-.l.\

o

Retraining epochs histogram for 2000 generation-old population.

75 100Retraining epochs

5025

O ..j.LLWlli1lJ.t1J=-c>£l.<"",",,,,",,,,,"",,,,,~~_ _ ....,... ~_..,-- ---.l.\

o

Figure 1: Re combination performance benchmark (continued).(b) P erformance benchmark for functionally matched recombination.

trained nets have good gener alization capabilities and additional steps-suchas pruning, weight decay, and so forth [19]-must be taken to improve gener­alizat ion. Generalization cap ability is a function of the capacity of the trainednet configuration and the complexity of the task itself [22, 23]. Training apowerful net to perform a simpl e task results in perfect training but poorgeneralization, while training an over-simplified net to perform a complicatedtask may result in an incomplet e training. Many generalization failures aredue to overfitting or "t uning to the noise," a situation where nets learn th enoise as well as the data. Pruning and other generalization enhancementtechniques help alleviate these problems by reducing net cap acity, usu ally byminimizing the number of free parameters in t he net.

This experiment demonstrates the generalization capabilit ies of the pro­posed algorithm and compares its performance with that of a computation-

Page 13: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Genetic Recombination 457

ally equivalent non-genetic training algorit hm. Genet ic evolut ion was carr iedout by the algorithm describ ed in sect ion 5, while non-genetic tra ining wasdone using randomly created nets instead of genet ically created offspring.We refer to the second meth od as selective back propagat ion . The randomlycreated nets had an initi al st ructure of 16:14:1 (fully forward connected) ,were t ra ined, pruned, and ret rained, and had to comp ete for a place in thepopulat ion. In both cases , the initi al population included the same 20 netsthat were pruned and retrained by the selected training dat a set , and th esame train ing and pruning parameters were used.

The task selected for th e demonstration consist ed of t ra ining a net toperform a Symmetry test for binary st rings, returning +1 if half th e st ringis a mirror reflection of the other half, and -1 otherwise. The input st ringswere 16 bits long, spanning an input space of 216 vectors. From this set, threedisjoint subsets were creat ed at random, each containing 200 vectors. Thefirst was used as a t raining set, the second for resolving the fitness of newlycreated and ret ra ined offspr ing , while t he last set was used as a measure ofthe generalizat ion ability of the nets . Note that t he last set was used onlyfor performance testing, and did not influence the training.

The results shown in Figure 2 give t he average fitness and genera lizat ionscores for all the net s in t he two populations, th e first having evolved bygenetic evolut ion and t he other by select ive back propagation. The averagefitness and generalizat ion scores are displayed as a function of the gener­at ion/ iteration count . The average t ra ining scores of both popu lations areapproximately 100% at all t imes and therefore not displayed. The superiorityof genet ic evolution is demonstrated by the improvement of both fitn ess andgeneralization scores throughout generations, while only minor improvementis found for select ive back propagation . After 1000 generations, genet ic evo­lution achieved a population t ra ining score average of 99.5%, a fitness scoreaverage of 91.5%, and a genera lizat ion score average of 90%. The best netachieved tra ining, fitn ess, and generalizat ion scores of 100%.

After 1000 iterations the select ive back propagation algorithm achieveda training score average of 99.5%, a fitness score average of 83.5%, and agenera lizat ion score average of 83%, while the best net achieved a t rainingscore of 98.5%, and fitness and generalization scores of 96.5%. The increase inboth fitness and genera lization scores and the results obtained by comparisonwith the selective back propagation method demonstrates the sup eriority ofgenet ic evolut ion and the success of its implementation for evolving neuralnet populat ions.

6.3 A ctivation function adaptat ion

We now demonstrate t he versatility of the proposed hybrid algorithm byshowcasing its ability to automatically choose an appropriate neuron acti­vation function. Populations of heterogeneous nets were evolved, each com­posed of a mixture of neurons where the activation function was randomlyselected from four different types of functions:

Page 14: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

458 Nachum Shamir , David Saad, and Emanu el Marom

Genetic evolutionSelective back propagation

Population fitness score average:

95%

85%~~-~--------,~----------,--.

90%

100%

750 1000

Generations/Iterations500250

80%+-- - - - - - ---,- - - - - - -----,,--- - - - - - --,-- - - - - - --,o

Populationgeneralizationscore average: Genetic evolutionSelectiveback propagation

85%

90%

95%

100%

750 1000Generations/Iterations

50025080%+-- - - ------,-------,----------,---------,

o

Figure 2: Generalization benchmark. Th e average fitn ess and gen­era lizati on scores of all 20 nets in the population are displayed asa function of the generat ion count for both a genet ically evolvingpopulati on of nets and a population evolving by selective back prop ­agation. Nets were tr ained by the Symmetr y-16 dat a set, where 200vectors were used for tr aining, 200 for resolving fitness of results, and200 for testi ng th e generalization capability of the nets. Th e averagetr ainin g scores of both populations was approximate ly 99.5% at allt imes and therefore not displayed. The superiority of genet ic evolu­tion is demonstrat ed by the marked improvement of both fitness andgeneralizat ion scores throughout th e generat ions , while only minorimpro vement is evident for th e selective back prop agation method.

• f( x ) = t anh(t x) , most commonly used for neural net training.

• f( x) = exp( - tx 2) , commonly used for classification t asks.

• f (x ) = xexp(- tx 2) , provides higher versatility t han t he exp(-tx2

)

function.

• f( x) = sin (tx) , whi ch separates any given set of point s on the real ax isusing a sing le par ameter.

The two tasks selecte d for the demonstration were exhaustive trainingby both t he Parity -6 and Symmetry-6 data sets . A population of 24 netswas randomly initialized and an act ivat ion function was chosen at random,

Page 15: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Genetic Recombination 459

Activation function distribution:

(Training by Parity-6 data set)

! (x) =tanh(tx)! (x) = exp(-tx 2)

! (x) = sin(tx)! (x) = x ' exp(- tx 2)

... ,.. . .- - - - - - - - - - - - - - --

c: 100%,:2:; 75%~1:l</)

50%is25%

250 500 750 1000

Generations

(a)

75%

50%

1000

Generations

! (x) = tanh(tx)!(x) = exp(- tx2)

! (x) = sin(tx)!(x) = x ' exp(-tx 2)

750500250

(Training by Symmetry-6 data set)

Activation function distribution :

c: 100%,:2:;.D'5

</)

is

(b)

Figure 3: Evolution of activat ion funct ion distribution. The distr i­bution of neuron types in the population is displayed as a functionof the generation count for the two exhaustive tr aining experimentsdone by (a) Parity-6 and (b) Symmetry-6 data sets, In the Paritytra ining exercise, neurons operating with t he sin(tx) activat ion func­t ion dominated the populat ion, compr ising 65% of the hidden andout put neurons in the ent ire populat ion after 1000 generations. Inthe Symmetry training example, the population was dominated byneurons operating wit h the tanh (tx) and exp(-tx2 ) activat ion func­tions, composing (after 1000 generations) 39% and 41% of the hiddenand output neurons, respectively.

thereby forming net s composed of a mixture of neuron types, T he selectionprocess was evenly dist ributed among all participating neuron typ es, resul t­ing in a uniform distribution of types in t he in iti al population, Each net inthe initial population was trained by back propagation and training was lim­ited to 100 epochs. The choice of activa t ion function is displ ayed in F igur e 3,

Page 16: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

460 Nachum Shamir, David Saad, and Emanuel Marom

Population average number of hidden neurons

(Training by Parity-6 data set)

10

7.5

5

2.5

1000

Generations

750500250ot----=:==:=:==~=======;========,o

(a)

Population average number of hidden neurons

(Training by Symmetry-6 data set)

10

7.5

5

2.5

1000

Generations

750500250o +------~------__,_------.__-----__,

o

(b)

Figure 4: Average number of hidden neurons in t he populat ion. Theaverage number of hidden neurons is displayed as a function of thegeneration count for t he two exhaust ive training experiments done by(a) Parity-6 and (b) Symmetry-6 data sets . In t he Parity tra iningexample, t he average was 0.67 hidden neurons after 1000 generations,indicating the existence of many perceptrons in the population. In t heSymmet ry tra ining example, the average number of hidden neuronswas 1.75 after 1000 generations.

t he average nu mber of hidden neurons in F igur e 4, and t he average t rain­ing scores are shown in Fi gure 5. All distribution curves (F igur e 3) start at25% because of t he uniform distribution of ac t iva t ion functions in t he ini ti alpopulat ion . As t he populat ion of nets evolved, t he distributio n of neurontypes, t he average number of hidden neur ons, and the average scores wereobserved for each generation. Since t he hybrid system is capable of adding

Page 17: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Genetic Recombination 461

Population average training score

(Training by Parity-6 data set)

100%

75%

50%

25%

0%0 250 500 750 1000

Generations

(a )

Population average trainingscore

(Training by Symmetry-6 data set)

100%

75%

50%

25%

0%0 250 500 750 1000

Generations

(b)

Figure 5: Average training score. In each of th e two experiments ,average t raining score increased as the population evolved. In bothcases, the best net performance was 100%, and the average scoreswere 95.3% for the Pa rity-6 training and 93.8% for th e Symmetr y-6training after 1000 generations.

and removing neurons, t he dist ributi on of activation fun ction ty pes changedover time, au tomat ica lly adapting toward optimal net st ruc t ure and neuroncomposit ion .

For the Parity training example (Figure 3(a)), neurons operating with thesin(tx ) act ivat ion func ti on gradua lly dominated the population, comprising65% of the hidden and ou tput neurons after 1000 generati ons. The averagesize of the net s decreased (Figure 4(a )) : after 1000 generations the averagewas 0.67 hidden neurons, indicating t he existe nce of many "percept rons" inthe p opulat ion .1 The average t ra ining score increased over t ime (Fi gure 5(a) ),

1A single node with a sinusoidal activation function representsa line that can separateany set of points in any given way using a single parameter; such a node possesses poor

Page 18: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

462 Nachum Shamir, David Saad, and Emanuel Marom

Average numberof hidden neurons(All hidden and output neurons operate with the tanh(tx ) activation function)---- Parity data set

Symmetry data set12

9

6

3

1000Generations

7505002500 +-- --- ------,--- ------,-- ----- ,-------- -,

o

Figure 6: Reference training. Average number of hidden neuronsis shown as a function of the generation count for two exhaust ivetraining experiments done by Parity-6 and Symmetry-6 data sets. Allnets in the populat ion contained hidden and output neurons operat ingwith the tanh(tx ) act ivat ion function. The average number of hiddenneurons after 1000generations was 5.2 for the Parity training set and3.6 for Symmetry.

reaching 95.3% after 1000 generations.For the Symmetry tr ain ing example (Figure 3(b)), the populat ion was

dominated by neuro ns operating with the tanh( tx) and exp(-tx2) act iva­

t ion functions, consist ing of (after 1000 generat ions) 39% and 41% of thehidden and output neurons, respectively. T he average size of the nets againdecrease d (Fig ure 4(b)): aft er 1000 generations t he average was 1.75 hiddenneurons. The average t ra ining score increased wit h each genera t ion (Fig­ure 5(b) ); reaching 93.8% after 1000 generations. In both expe riments , thebest net perform ance was 100%.

The size of the nets for both the Parity an d Symmet ry t ra ining experi­ments was significant ly smaller than the sizes required for nets where all hid­den and output neurons operate with the tanh( tx) activation function. T heresul ts displayed in Figure 6 show th e average number of hidden neurons forall nets in a population utilizing the t anh(tx) activat ion func tio n, exhaus­t ively t ra ined by the Pari ty-6 and Symmetry-6 data sets . The experimentwas performed using the same para meters as in the previous experiments.The average number of hidden neurons aft er 1000 generations was 5.2 for theParity training set and 3.6 for Symmetry.

The results of these experiments demonstrat e the capability of the pro­posed algorithm to minimize net st ructure , choos ing t he most suitable ar­chitect ure and act ivat ion funct ion for a given task. T he versat ility of t he

generalization capabilities. Wh en trai ning is done for gener alizati on purposes, as it isin most cases, one should exclude sinusoidal decision functions to improve general izat ioncapabilit ies.

Page 19: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Genetic Recombin ation 463

m 2~'-l567f! "1 01 2.3<Cf.,&7S'l CH 2dlfS"67lJ? or 23'tSb7lt't01 21l~'5r;;7<!i'9 DI '7..3<fS-'::;7Bo<f O/2.3"fr6?l! 1' orZ3<jost;,7~~

01 2~"'5l:.1""~ 01 '7.3"fS"b78'" (}/2iJ~r67li' 'P Of .z3'f"~78qor 21>~"'b 7 8"" 01 2.3"i'5678'1' ()/ ;;'lYfjtf7e'i' 0IZ3'f-S'-78"iat 2.3 " S EJ 7~'" 01 234S"679., CU ;;'.3(i..rtf ? ll' ~ 01 2-3'f-S"b7a-901 23.Lt",r; 7g Of Cll "-345'"67:B"l t')/;;;".3~r6.?lt~ OIZ..34-~"78'';lI01 2"..4'::::iEJj'l!"t m 2.3q.S£.78'9' Q r:2~~J"'tf?ltl' 0 1 2.3,+r;"7~91)[ 2.3"1 SIS"1 El9 or .23q.~{;7B"t () / 2.d r.L S" 6?R' 'ii' 0 1 2.3'+'1>'-78'901 2345b-''il.~ 01 23"'~ '7;!'l"t {} /,2d~.r~7R" e>1 2..3'fS""-78"'fOJ 2~/';'::::i~7El'l OL 2.3 'H· "'7~"'l "/L~ fftS ~'? 1i'? 01 Z-3r.t""7lr'l'

01 'l?4St;'l?q 01 23If-S&7fi:'l 01 2~ 'is l.Til'f (")1 2.3 '1 S"r:;795'(~H 2.~CfS"6. ,Siq 01 Z34.567i?"7 01 'J."l'f5~7 fl'1 OI23'1~~7'i1 "01 2..Yi~b ,5<'<;1 0'2~'l-S67~'7 01 z J vsc 7fl "l 0 1 23't'5"&;7B701 2.~'f"6Igq 01 2.3't-567Ii'~ m '23'f~€.7fl '1 012.3'1'S'{;7'8",/or 2.~lI-~~j~'t 01 23'f.3&7117 ('JI as Ifrd . 7 t!., Q/23"1S"C7B'T01 2~q.'b7B.<J OI2~'+Sf>7.!J':I Dl 2~'I!i("72'l O'Z3"1S"r:;78»O! 2.~Q"~78't 012.3"-567117 CI 2:J'fS1.7117 ot 23-'Y~{;7~~m 2.3q.-,(,,/~'j 01234.5(,,73'1 01 :23"',5"1:: »er en 2Y/J~7'jj't'Of 2:!,,(fr:l:.7R'1 (J/ .2..3l.j.S " 7lJ 'l OI2;lyS"7tlY ot 23'15"&.7~?01 2~4.s-6/li'q 01 Z3~567k"1 CI23'fSli.7N'l' 01 2.3'1" J ~7lJ'''

or 23<t5~7"lf" Ol2"!: '1Hi7 'BCI 01 .;;I,34!;"c,,7<;lQ QI2~(,l5"' '1:S9(J(2"3'f S"-'''B'9 Of ;U'1~&.7 lf'! m 23'i5"'7~q OI.2..31J6(",i9Q('.J (2:3'1~157if9 O/J.~ "i~67lJ 9 01 23'iS,"7'1!i'1 O'2..3o(15~"89O/2"34,!j~7lf~ eli 23 '1rl;,] 1l9 01 ;l.3"l5C:;7~,) {J12 ..'J ""5"~7 29'Of'22l ..... J"t.7Pc;> f::j':l:i ~lt'~78" 01 ;;l. '3 '-lS ,- j 'il"t 0/234£5""18'11'(j/Z~""'S"7T" 01 ~3"1!:" &.n/' 'i m ,~,:3'i5 '" 7'il."l 0 1 23(J.!>",.,8<fa '7..3"f S 67 K"J 0 1 il.3'1 t'€.711'i 01 ;1. 3 '-\5 " 7 'il:'1 01 2 3«1S& 7,.. "IQI 2. ~"'51G. 7k"i' O f 23 'f ~"7 1/" Ol ;1.. 3 'i S," 7';l;<J C / Z.34!J{,7 '8"1Cd 2 ,!;'I<;' 678 ~ Of:J.3'1s'E: 7 lf 'l 0 1 :;)' 3'iS"7~q DI2 .34S-"'789CJt :J.~ Y!i"7lf" 01 :/3'fr cr e» Ol :l3'-i5 (;'j"i!:q DI.l..3Q S "'7Sl''T

Figure 7: Handwritt en digits. This set of 1200 digits was writtenby twelve people, ten times each. Due to computational limitat ions,data resolution was reduced from the original sampling resolution of16 x 16 pixels to 8 x 8 pixels for each digit.

proposed algorit hm is therefore increased beyond a simple const ruct ion andtrain ing algorit hm, consisting of a method for adjusting intern al neuron pa­rameters (decision functions) t hroughout the pro cess of t ra ining.

6 .4 Handwritten digit s recogni tion

The t ask of identifying handwritten digits provides a prac t ical example of theproposed training syste m and highlights its ability to automatically adapt todiverse tas ks. The dat a base includes 1200 digits writ ten by twelve differentpeople, ten times each (the dat a base was prepared by 1. Guyon) . Dueto computationa l limitations, dat a resolution was redu ced from the originalsampling resolution of 16 x 16 pixels to a resolution of 8 x 8 pixels for eachdigit. The dat a set is displayed in Figure 7.

Training was divided into ten different runs, each dedicated to producinga single net capable of identifying a particular digit and rejecting all ot hers .The origina l data set was split at random into two disjoint sets of 600 digitseach. The first set was used for t ra ining and fitness evalua t ion and th esecond set was used to evalua te th e quality of the final results by measuringthe genera lizat ion ab ility of the fabricated nets. All t rainings used the sameparameters. The init ial net st ructure was 64:6:1 (fully forward connected)

Page 20: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

464 Nachum Shamir, David Seed, and Emanuel Marom

Digits identi ficat ion results

Structure ResultsNumb er of hidden Number of Train Test Errors

Digit neurons interconnect ions Success Success

0 0 8 100.0% 99.2% 5

1 0 7 100.0% 98.0% 12

2 1 24 100.0% 98.3% 10

3 2 24 100.0% 97.7% 14

4 0 15 100.0% 97.8% 13

5 1 19 100.0% 97.7% 14

6 1 19 100.0% 97.8% 13

7 1 15 100.0% 99.7% 2

8 1 32 100.0% 96.7% 20

9 3 150 100.0% 96.3% 22

Table 1: Result of digit-identification experiment . The st ructures andperformance are presented for each of the 10 nets. All nets performedwith a 100% success rate on the 600 training digits, and had extremelyhigh test scoreson the remaining 600test digits. The resulting nets arevery small; in particular, the nets for digits 0, 1, and 4 are perceptrons(with no hidden neurons).

where hidden and out put neurons operated with a tanh(x) decision function.Back propagation initial training, ret raining after genet ic recombin at ion, andret raining afte r pruning were limited to 100 epochs. The pop ulation size foreach t ra ining was 10 nets. The init ial population was created by choosingthe best results from 50 random trainings, which we th en let evolve for 200generations.

The result s are summ ar ized in Table 1, which describ es th e dimension ,final training score (ident ical with the fitness score), and the test perform ancefor the ent ire digit set of each of th e ten nets. All nets performed with a100% success rate on the 600 tr aining vectors and had ext remely high test ingscores for the remaining 600 vectors. The result ing nets were very small. Inpart icular , t he nets t ra ined to recognize digits 0, 1, and 4 are percept rons(with no hidden neurons).

The results in Tab le 1 show the performance of each net t rained to ac­cept a single digit and reject all others . One identifies the actua l digit byfeeding the same dat a to each of the ten nets and selecting t he net givingt he highest out put value. For the ent ire set of 1200 digits, the number ofcorrect identifications was 1160, which is 96.7% of t he ent ire data set. Gen­era lization performa nce was calculated for the 600 digits that were not usedfor t ra ining, giving a success rate of 93.3%. One should note that no param­eter optimization was carried out and no special net st ruct ure was created

Page 21: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Genetic Recombination 465

prior to tra ining. Thus, all st ruc tural refinements were automa tically doneby genet ic t ra ining and evolut ion rules.

7. Conclus ion

In this paper, we propose a new hybr id genet ic back propagat ion t ra ining al­gorithm based on a unique functional matching recombin ation method. Themethod is used to evolve heterogeneous populations of neur al networks andprovides versatility in network architec ture and activation functions. Theperforman ce of the proposed algorit hm was tested on a wide variety of experi­ments, showing its ability to overcome th e problems originat ing from locat ionpermutations of hidden neurons and to efficient ly hand le heterogeneous pop­ulat ions of nets having diverse st ructure and functional composit ion. Vit alinfrast ructures are preserved throughout recombination and transferred tothe next generation, thereby impro ving the quality and initial performanceof the generate d offspr ing.

We performe d four experiments t hat demonstrat e th e utility of th e pro­posed hybr id. T he first experiment demonstrates the importance of func­tional reorganization of nets duri ng genetic recombinat ion, showing t hat vitalinfrastructures are transferred from parent to offspring nets only when func­tional reorganizat ion is carr ied out. Functional reorganization was thereforefound to have a critical influence on the success of genet ic implementationand the ability to preserve and improve "genet ic charact erist ics" throughoutthe evolut ionary pro cess.

In the second experiment, the genera lizat ion propert ies of the proposedhybrid algorithm were tested using the Pari ty-16 data set , and the resultswere compared to a computat ionally equivalent non-genetic tra ining example,showing indispu t ably the advantage of genetic evolut ion by producing moreefficient nets with higher generalizat ion capabilit ies. T he third experimentdemonst rated the ability of the hybrid to handle a populat ion of nets withheterogeneous functional composition, dynamically adapting both th e st ruc­ture and composit ion of the populat ion . In t he last experiment, we tra inednets to ident ify handwritten digits where all struct ural refinements were au­tomat ically done by genet ic training and evolut ion rules. These experimentsdemonstrate the efficiency and success of the implementat ion, highlightingthe enormous contribut ion of genetic search to the success and adaptabilityof neur al network t raining.

Acknowledgments

We th ank J. Shamir for helpful discussions, advic e, and supp ort, and to1. Guyon for providing the data set for the handwritten digits.

R eferences

[l] N. Sharnir , D. Saad , and E. Marom, "Neur al Net Pruning Based on Fun ctionalBehavior of Neurons," International Journal of Neural Systems, 4 (1993) 143­158.

Page 22: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

466 Nachum Shamir, David Saad, and Emanuel Marom

[2] N. Shamir , D. Saad , and E. Marom, "Preserving th e Diversity of a Geneti­cally Evolving Populati on of Nets Using the Functional Behavior of Neuron s,"Complex Systems (in press, 1994).

[3] D. E. Rumelhart , G. E. Hinton, and R. J. Williams, "Learn ing Internal Rep­resentations by Error P ropagation ," pages 318-362 in Parallel DistributedProcessing: Explorations in the Microstructure of Cognition , edited by D. E.Rum elhar t and J. L. McClelland (Cambr idge: MIT Pr ess, 1986).

[4] J. Sietsma and R. J. F. Dow, "Neural Net Pruningv-W hy and How?" pages325-332 in Proceedings of the IEEE International Conference on Neural Net­works, 1, San Diego, CA (1988).

[5] M. C. Mozer and P. Smolensky, "Skeletoni zat ion: A Techniqu e for Trimmingthe Fat from a Network via Relevance Assessment ," pages 107- 115 in Ad­vances in Neural Inform ation Processing, 1, edited by D. S. Tour etzky (SanMateo, CA : Morgan Kaufmann, 1989).

[6] E . D. Karnin , "A Simple Pr ocedure for Pruning Back-P ropagat ion TrainedNeural Networks," IEEE Transactions On Neural Networks, 1 (1990) 239­242.

[7] M. Mezard and J. P. Nadal, "Learn ing in Feedforward Layered Networks:Th e T iling Algorit hm ," Journal Physics A , 22 (1989) 2191-2203.

[8] M. Frean , "The Upst art Algorithm: A Met hod for Const ructin g and TrainingFeedforward Neural etworks," Neural Computation, 2 (1990) 198- 209.

[9] S. E. Fahlm an and C. Lebiere, "T he Cascade-Correlat ion Learn ing Architec­ture," pages 524-53 2 in Advances in Neural Inform ation Processing Systems,2, edited by D. S. Touretzky (San Mateo, CA: Morgan Kaufmann, 1990).

[10] S. E. Fahlman, "The Recur rent Cascade-Correlation Archit ecture," pages190-1 96 in Advances in Neural Information Processing Systems, 3, editedby R. P. Lippmann , J. E. Moody, and D. S. Tour etzky (San Mateo , CA :Morgan Kaufman n, 1991).

[11] D. E. Goldberg, "Genet ic Algorithms in Search, Optimizat ion , and MachineLearning" (Reading, MA: Addison-Wesley, 1989).

[12] D. Wh it ley and T . Hanson, "T he GE NITOR Algorithm: Using Geneti c Re­combination to Optimize Neural Networks," Technical Report CS-89-107, De­par tment of Comput er Science, Colorado State University (1989).

[13] D. Whi tl ey and T . St arkweather, "GENITORII : A Dist ribu t ed Genet ic Algo­rithm," Journal of Experim ental Theoretical Artificial Intelligence, 2 (1990)189- 214.

[14] D. Wh itl ey, T . Starkweath er , and C. Bogar t "Genet ic Algorithms and NeuralNetworks: Optimizing Connect ions and Connect ivity," Parallel Computing,14 (1990) 347-361.

Page 23: Using the Functional Behavior of Neurons for Genetic Recombination in Neural Nets Training · 2018-03-14 · Using the Functional Behavior of Neurons for Genetic Recombination in

Using the Functional Behavior of Neurons for Genetic Recombination 467

[15] H. Kitano "Designing Neural Networks Using Genet ic Algorithms wit h GraphGenera tion Syst em," Compl ex Systems, 4 (1990) 461-476.

[16] D. J . Mont ana and L. Davis, "Training Feedforward Networks Using GeneticAlgorithms ," pages 762-767 in Eleventh Inte rnational Joint Conference onArtificial Intelligence (Detroit 1989) , edited by N. S. Sridharan (San Mateo,CA: Morgan Kaufmann, 1989) .

[17] Y. Le Cun, J . S. Denker , and S. A. Solla , "Opt imal Brain Damage," pages598-605 in Advances in Neu ral Info rmation Processing Systems, 2, edite d byD. S. Touretzky (San Mateo, CA: Morgan Ka ufmann, 1990).

[18] B. Hassibi, D. G. Stork, and G. J. Wolff, "Optimal Brain Surgeon and Gen­era l Network Pruning," pages 293-299 in IEEE Intern ational Conference onNeural Netwo rks, San Francisco (P iscat away, NJ : IEEE Press , 1993).

[19] A. S. Weigend, D. E. Rumelhart, and B. A. Huberman, "Generalizat ion byWeight -Elimination with App lication to Forecasting," pages 875-883 in Ad­vances in Neural Information Processing Systems, 3, edited by R. P. Lipp­mann, J. E. Moody, and D. S. Touretzky (San Mateo , CA: Morgan Kaufmann ,1991) .

[20] N. Radcliffe, "Geneti c Neural Networks on MIMD Machines ," Ph.D. Thesis,University of Edinburgh (1990).

[21] N. J . Radcliffe, "Genetic Set Recombination and its App lication to Neur alNetwork Topo logy Optimization," Neural Computing fj Application, 1 (1993)67-90.

[22] V. N. Vapn ik and A. Y. Chervonenkis, "On the Uniform Convergence ofRelative Frequ encies of Event s to Their Probabilities," Theory of Probabilityand Its Applications , 16 (1971) 264-280.

[23] V. N. Vapnik , "Est imat ion of Dependences Based on Empirical Data,"(Berl in : Spr inger-Verlag , 1981).