Upload
harinima
View
217
Download
0
Embed Size (px)
Citation preview
8/9/2019 theis4
1/27
124
CHAPTER 6
HYBRID METHODS FOR MINING ASSOCIATION RULES
6.1 GENERAL
This chapter reviews the necessity of hybridizing GA and PSO methods,
salient aspects of GA and PSO hybrid methods as available in published
literature, and proposes a new hybrid methodology of GA and PSO for AR !
As the main drawbac" of PSO being wea" local search, it is overcome by
proposing Shuffle #rog $eaping Algorithm for the local search to mine ARs!
The e%perimental results of both proposals are also presented, and discussed
along with the salient inferences there from!
6.2 NEED FOR HYBRIDIZATION
The PSO was inspired by insect swarms and has been proven as a
competitor to the standard GA for function optimization! Since then several
researchers have analyzed the performance of the PSO with different settings,
e!g!, neighborhood settings &'ennedy()))* Suganthan()))+! omparisons
between PSOs and the standard GA were done analytically in &-berhart and
Shi()).+ and also with regard to performance in &Angeline ()).+! /t has been
pointed out that the PSO performs well in the early iterations, but has problems
reaching a near optimal solution in several real0valued function optimization
problems &Angeline ()).+! 1oth -berhart and Angeline conclude that hybrid
models of standard GA and PSO could lead to further advances!
6.3 HYBRID OF GA AND PSO FOR ARM
AR using either GA or PSO methodology often generates rules with
high PA! 1ut balancing between e%ploration and e%ploitation is often a
tailbac"! GA has good e%ploration capability, but at times it leads to e%ploiting
the search space! PSO has the capability to converge 2uic"ly thus avoiding
8/9/2019 theis4
2/27
125
e%ploitation but results in premature convergence! 3ybrid system combining
GA and PSO is a solution for the balancing phenomena!
Recently the hybridization between evolutionary algorithms &-As+ andother metaheuristics has shown very good performance in many "inds of multi0
ob4ective optimization problems, and thus has attracted considerable attention
from both academic and industrial communities!
An evolutionary circle detection method based on a novel haotic
3ybrid Algorithm combines the strength of PSO, GA and chaotic dynamics! /t
involves the standard velocity and position update rules of PSO, with the ideas
of selection, crossover and mutation from GA & hun03o at al!, 56(6+! A hybrid
multi0ob4ective evolutionary algorithm incorporating the concepts of personal
best and global best in PSO and multiple crossover operators to update the
population, maintains a nondominated archive of personal best &Tang and
7ang, 56(5+! A hybrid method combining GA and PSO creates individuals in a
new generation not only by crossover and mutation operations as found in GA,
but also by mechanisms of PSO! #urther the above approach solves the
problem of local minimum of the PSO, and has higher efficiency of searching
global space & 8ie Ru and 9ue :ianhua 566.+!
The hybridization methods of GA and PSO analyzed here are process
based, integrating the GA steps into PSO or vice versa! Population based
hybridization is attempted in this study, where the population is split based onfitness values and both methods are run individually on the respective
subpopulation! The ob4ective of this hybrid methodology is to sum up the
advantages of both GA and PSO in overcoming the drawbac"s!
6.3.1 GA-PSO Hybri A!"#ri$%& '#r ARM
8/9/2019 theis4
3/27
126
-ach evolutionary algorithm proposed in the literature has some
advantage over the other! PSO provides a faster convergence than GA! A hybrid
techni2ue utilizing the effectiveness and uni2ueness of these two algorithms
can be implemented to achieve high performance &;ong et al!, 56(5+! The proposed hybrid methodology that is the GPSO proposed combines the features
of GA and PSO to enhance the performance of AR ! The uni2ueness of GA
and PSO are
• The optimization process of GA is a steady state process converging at
global optima!• The global search space is maintained by GA avoiding premature
convergence!• PSO is faster a process with the ability to search the solution space
2uic"ly• PSO "eeps trac" of each particles< best position thereby effectively
fi%ing up the search space
The hybrid methodology employs these features for evolution of individualsover generations! The total population is split into two subpopulations based on
fitness values! The fitness function designed is for ma%imization and the
individuals with ma%imum fitness values are aimed to evolve towards the
global optima! So GA is applied on the subpopulation with higher fitness
values! PSO with the ability to converge 2uic"ly is applied on the lower ran"ed
subpopulation based on fitness values! The GA with optimization at global
optima enhances the e%ploration capability and PSO with easy convergence
avoids e%ploitation! Thus the hybrid methodology effectively balances between
e%ploration and e%ploitation thereby resulting in ARs with better PA! The upper
and lower ran"ed subpopulations after evolution through GA and PSO are
updated at the end of each iteration! This aims at consistency in performance,
avoiding the loss of the better individuals over generations!
8/9/2019 theis4
4/27
Genetic Algorithm
Particle Swarm Optimization
Evaluate Fitness
Upper
Lower
Initial Population Ranked Population Updated Population
127
Applying GA on the upper ran"ed subpopulation avoids premature
convergence! -volution through PSO on lower ran"ed subpopulation avoids
e%ploitation of GA of individuals, thereby increasing the speed of evolution
process! The proposed hybrid model is shown in #igure =!(!
Fi"(r) 6.1 Hybri GA* PSO +GPSO, M# )!
As can be seen in #igure =!(, GA and PSO both wor" with the same
initial population! The hybrid approach ta"es 8 individuals that are randomly
generated! These individuals may be regarded as chromosomes in the case of
GA, or as particles in the case of PSO! The 8 individuals are sorted by fitness,
and the upper half individuals are fed into the GA to create new individuals by
crossover and mutation operations! The lower half of the population sorted
through fitness values are fed into PSO for evolution! The velocity update and
position update are done based on the personal best and global best positions
determined on the lower half individuals! The output of GA and PSO are
combined for propagating into ne%t generation and the process repeated till an
end criterion is met!
8/9/2019 theis4
5/27
128
1y performing evolution based on GA on the upper half of the individuals
sorted through fitness value the global optimal solution space is retained! This
facilitates the steady progress towards the optimization through e%ploration!
PSO applied on the other half of the individuals where fitness value is low, theconvergence is achieved easily avoiding e%ploitation! Thus, the hybrid GPSO
model balances between e%ploration and e%ploitation by combining the
strengths of GA and PSO! The pseudo0code for GPSO is given below>
/nitialize all GA variables
/nitialize all PSO variables
Repeat
alculate fitness value of the population
Ran" the population based on fitness function
Split the population into two halves> higher ran"ed range, lower
ran"ed range
On higher ran"ed range partition perform GA
S)!) $i#/0 Select two parent chromosomes from the population
according to their fitness
Cr# # )r0 7ith a crossover probability cross over the parents to
form a new offspring
M($ $i#/0 7ith a mutation probability mutate new offspring at
each locus
U4 $i#/0 Place the resulting new offspring in a new population
On low ran"ed partition perform PSO
#or each particle
U4 $) velocity
U4 $) position
8/9/2019 theis4
6/27
129
U4 $) p1est and g1est
-nd for
?pdate new population combining PSO particles and GA chromosomes
?ntil &stopping condition+
The population size is fi%ed based on the size of the dataset for which
AR is applied! 1inary encoding is adopted for representation of data! The
fitness function as given in -2uation &@!5+ is adopted for calculating the fitness
values! The e%perimental setting and results of the GPSO methodology for
mining ARs are presented in ne%t section!
6.3.2 E54)ri&)/$ ! R) (!$ / Di ( i#/
To test the performance of the hybrid GPSO for mining ARs,
e%periments were carried out on the well0"nown benchmar" datasets from ? /
repository!
The parameters, which play a ma4or role during the rule discovery in the
hybrid GPSO methodology, are listed in Table =!(! The population size is the
size of the individuals ta"en up for e%perimentation! The crossover and
mutation rates are the GA operator specifications! c ( and c 5 are the acceleration
coefficients used in velocity updation of PSO as in -2uation & !5+!
-volutionary algorithms are relatively simple to implement, robust and
perform very well on a wide spectrum of problems! This study proposes a
hybrid methodology of evolutionary algorithms> GA and PSO for AR ! The
scope of this study on mining ARs using GPSO is to>
• Study the performance of GA over generations• Analyze the performance of PSO over generations• /dentify the limitations of GA and PSO while mining ARs in terms of
PA and e%ecution speed!• ompare the performance of GPSO with GA and PSO
8/9/2019 theis4
7/27
130
T b!) 6.1 GPSO P r &)$)r '#r ARM
P r &)$)r !()Population Size $enses > 56
ar -valuation > B663aberman 66Post0operative Patient are > .6Coo > (66
rossover Rate $enses > 6!=ar -valuation > 6!B
3aberman 6!BD
Post0operative Patient are > 6!.Coo > 6!.utation Rate $enses > 6!D
ar -valuation > 6!@3aberman 6!5DPost0operative Patient are > 6!5Coo > 6!5
Selection Operation Roulette wheel selectionc( 5c5 5
8o! of Generations D6
E #!($i#/ # )r G)/)r $i#/ A/ !y i
GA is "nown for maintaining the global optima throughout evolution
and steady progress in performance over generations, whereas, PSO converges
2uic"ly with chances of converging at local optima! So to analyze the
performance of GPSO over generations, the ma%imum PA of the ARs mined byGPSO is recorded at intervals over evolution for all the five datasets! This data
is plotted against the results obtained with GA and PSO as shown in #igure =!5!
8/9/2019 theis4
8/27
131
&a+
GA0
10
20
30
40
50
6070
80
90
100
10
20
30
40
50
ethodolog!
Predictive Accurac! "#$
8/9/2019 theis4
9/27
132
GA0
10
20
30
40
50
6070
80
90
100
10
20
30
40
50
ethodolog!
Predictive Accurca! "#$
0
10
20
30
40
50
60
70
80
90
100
10 20 30 40 5 0
ethodolog!
Predictive Accurac! "#$
8/9/2019 theis4
10/27
133
0
10
20
30
40
50
6070
80
90
100
10 20 30 40 50
ethodolog!
Predictive Accurac! "#$
ar -valuation ;ataset
&b+ 3aberman
8/9/2019 theis4
11/27
134
&e+ Coo ;atasetFi"(r) 6.2 Pr) i $i ) (r y . N#. #' G)/)r $i#/
'#r ARM 7i$% GPSO
#rom #igure =!5 it is observed that mining ARs using GA results in
lesser PA than PSO and GPSO! The increase in PA is achieved steadily over
generations! The PSO methodology for mining ARs generates AR with better
PA at earlier stages of evolution! /n further generations, the particles move
away from global optima bringing down the accuracy, thus e%ploiting the
search space! #or all the five datasets, GPSO methodology generates ARs with
better PA and maintains the same over generations, thereby balancing between
e%ploitation and e%ploration! Thus the stability in performance is obtained
while mining ARs with GPSO!
#rom the #igure =!5 a to e it is observed that• The performance of GA in terms of PA increases over generations
indicating its effectiveness in terms of global search capability*• PSO generates ma%imum accuracy at initial generations thereby
converging 2uic"ly*• The deviation from global optima and convergence at local optima is
attained in later generation for PSO! This e%ploitation of search space
results in reduction of PA*• GPSO produces consistent PA with minimal difference over generations
and• The PA of GPSO is better than GA and PSO!
Pr) i $i ) A (r y A/ !y i
8/9/2019 theis4
12/27
135
The ob4ective of the study is to enhance the PA of ARs mined by
utilizing the uni2ueness of both GA and PSO! The PA of the ARs mined by
GPSO is plotted for the five datasets as shown in #igure =! ! The results
obtained using simple GA and PSO are also shown in same figure for comparison!
50
60
70
80
90
100
GA PSO GPSO
%ataset
Predictive Accurac!"#$
Fi"(r) 6.3 C#&4 ri #/ #' Pr) i $i ) A (r y #' GPSO 7i$% GA / PSO
The PA of rules mined with PSO is better than GA! The PA obtained by
GPSO is enhanced when compared to GA or PSO! The PA for the $enses and
Coo datasets are e2uivalent to that of PSO, but better than GA! /ncrease in PA
upto DE is achieved over PSO by GPSO! An increase of 6E is obtained by
GPSO over GA for 3aberman
8/9/2019 theis4
13/27
136
A/ !y i #' E5) ($i#/ Ti&)
AR with GPSO performs both GA and PSO operations on their respective subpopulation! The effect of GPSO on e%ecution time while mining
ARs is shown in #igure =!@! The e%ecution time of GPSO is better than genetic
algorithm and ta"es more time than that of PSO! The evolution of the lower
ran"ed population by PSO ma"es the convergence 2uic"er, thereby reducing
the e%ecution time achieved compared to GA! The GA operations ta"e more
time and are comple%, when compared to PSO! Thus there is an increase in
e%ecution time of GPSO over PSO! The increase in prediction accuracy over
PSO compensates the time trade0off!
As standalone both GA and PSO produce results inconsistently for all
the five datasets! The drawbac" of GA is its lac" of memory, which limits, its
search and convergence ability! The mutation operation also leads to
e%ploitation and hence the inconsistency in accuracy for all the five datasets!
PSO tends to converge at local optima resulting in premature convergence,thereby generating inconsistent results! 1y combining the advantages of both
GA and PSO the GPSO method mine ARs with consistent performance! The
GPSO method of mining AR outperforms both GA and PSO methodology in
terms of prediction accuracy, consistence and e%ecution time!
8/9/2019 theis4
14/27
137
&'&&&
(&&&
)&&&
*&&&
+&&&&
GA
PSO
GPSO
%atasets
E,ecution -ime "ms$
Fi"(r) 6.8 C#&4 ri #/ #' E5) ($i#/ Ti&) #' GPSO 7i$% GA / PSO
6.8 PSO 9ITH LOCAL SEARCH
Some important situations that often occur in PSO is overshooting ,
which is a "ey issue to premature convergence and essential to the performance
of PSO! #rom the velocity update mechanism of PSO, it is observed that the
p1 est and g1 best ma"e the particles oscillate! The overshooting problem
occurs due to the velocity update mechanism, leading the particles to the wrong
or opposite directions against the direction to the global optimum! As a
conse2uence, the pace of convergence of the whole swarm to the global
optimum slows down! One possible way to prevent the overshooting problem
from happening is to appropriately ad4ust the algorithmic parameters of PSO!
3owever, it is a difficult tas", as the parameter ad4ustment depends a lot on
domain "nowledge and the optimization problem!
/n conse2uence of overshooting, the particle will move to the opposite
direction against the direction to the global optimum! Two ma4or approaches
that can be used to tac"le the overshooting problem are described below!
8/9/2019 theis4
15/27
138
• A/ !y i #' 4r#b!)& > 1y analyzing the structure of problems or
identifying the fitness landscape of problems, a lethal movement could
be prevented! /t poses great challenge on automated problem structure
analysis &1ucci and Pollac", 566@+
• G)/)r $) / $) $ $r $)"i) > There are many methods to generate a
new solution for testing, such as> heuristics of the specific problem,
statistic sampling, and local search techni2ues! 3owever, the
computation costs in generating and testing new solutions may be
generally high!
/n order to develop a general0purpose algorithm and overcome the
overshooting problem, an efficient local search strategy F Shuffle #rog $eaping
Algorithm &S#$A+, is adopted and combined with the standard PSO for AR !
6.: PSO 9ITH SFLA FOR MINING ASSOCIATION RULES
Population0based heuristics inherently improve the implementation of a
local search algorithm, since the heuristic approach of a population of solutionsresults in rather poor local search properties! /ncorporating local search
algorithm into the population based heuristic is called a emetic Algorithm!
/n published literature many $S schemes have been employed with PSO
for optimization! Petalas et al! &566B+ employed a stochastic iterative $S
techni2ue in their A, called R7;-, where a se2uence of appro%imations of
the optimizer are generated by assuming a random vector as a search velocity!
/t was noticed by ictoire and :eya"umar &566@+ that early on in the PSO
search, particles were almost close to the pro%imity of the global optimum, then
move away from these areas! #or this reason the local search method was
chosen for implementation! /nspired from literature a memetic PSO with S#$A
for local search is proposed for mining ARs!
8/9/2019 theis4
16/27
139
The S#$A is a memetic metaheuristics that is designed to see" a global
optimal solution by performing a heuristic search! /t is based on the evolution
of memes carried by individuals and a global e%change of information among
the population &-usuff and $ansey 566 +! The S#$A involves a population of possible solutions defined by a set of frogs &i!e! solutions+ that is partitioned
into subsets referred as memeple%es! The different memeple%es are considered
as different cultures of frogs, each performing a local search! 7ithin each
memeple%, the individual frogs hold ideas that can be influenced by the ideas
of other frogs, and evolve through a process of memetic evolution! After a
number of memetic evolution steps, ideas are passed among memeple%es in a
shuffling process &$iong and Ati2uzzaman 566@+! The local search and the
shuffling processes continue until convergence criteria are satisfied &-usuff and
$ansey 566 +!
6.:.1 M)$%# #!#"y
/nitially the particles are distributed in the search space and fitness of the
particles is calculated! The velocity and position updation of the particles are
carried out using PSO
8/9/2019 theis4
17/27
G)/)r $i#/ #' i/i$i !4#4(! $i#/+P, / ) !( $i/" $%)
'i$/) #' ) % 4 r$i !)
)!# i$y / 4# i$i#/ (4 $i#/#' 4 r$i !)
Di $rib($i#/ #' 'r#" i/$# M&)&)4!)5)
I$)r $i ) U4 $i/" #' 7#r $ 'r#"i/ ) % &)&)4!)5)
C#&bi/i/" !! 'r#" $# '#r& /)7 4#4(! $i#/
T)r&i/ $i#/ ri$)ri
$i 'i) ;
D)$)r&i/) $%) b) $ #!($i#/
S#r$i/" $%) 4#4(! $i#/ i/) )/ i/" #r )r i/ $)r& #'
'i$/) !()
SFLA
140
Fi"(r) 6.: F!#7 % r$ '#r PSO 7i$% SFLA '#r ARM
/n S#$A, the population consists of a set of frogs &solutions+ that is
partitioned into subsets and it is named as memeple%es! The different
memeple%es are considered as different cultures of frogs, each performing a
local search! 7ithin each memeple%, the individual frogs have different ideas,that can be influenced by the ideas of other frogs, and evolve through a process
of memetic evolution! After a defined number of memetic evolution steps,
ideas are passed among memeple%es in a shuffling process! The local search
and the shuffling processes continue until defined convergence criteria are
satisfied!
The S#$A starts with an initial population of P frogs created randomly!
Then, the frogs are sorted in a descending order of fitness! Then, the entire
N
Y
8/9/2019 theis4
18/27
Frog 6
Memeplex
1Memeplex
2Memeplex
3
Frog 1
Frog 2
Frog 3
Frog 7
Frog 5
Frog 4
Frog 8
141
population is divided into m memeple%es, each containing 8 frogs! /n this
process, the first frog goes to the first memeple%, the second frog goes to the
Second memeple%, frog goes to the th memeple%, and frog H( goes bac"
to the first memeple%, etc! The process is as shown in #igure =!=! 7ithin eachmemeple%, the frogs with the best and the worst fitness are identified as I b and
I w, respectively! Also, the frog with the global best fitness is identified as I g!
Then, a process similar to PSO is applied to improve only the frog with the
worst fitness ¬ all frogs+ in each cycle!
Fi"(r) 6.6 F#r& $i#/ #' M)&)4!)5)
Accordingly, the position of the frog with the worst fitness is ad4usted based on
-2uations &=!(+ and &=!5+!
Change ∈frog position ( Di)= rand ()∗ X b− X w &=!(+
New Position ( X w )= Current Position ( X w )+ ( D i ) &=!5+
7here rand & + is a random number between 6 and (* X
b is the
position of best frog in the group* X
w
is the position of worst frog in the
8/9/2019 theis4
19/27
142
group! /f this process produces a better solution, it replaces the worst frog!
Otherwise, the calculations in -2uations are repeated, but with respect to the
global best frog &i!e! I g replaces I b+!
The fitness function defined in e2uation &@!5+ is used for evaluating the
fitness of the individuals! 1oth PSO and APSO methods are combined with
S#$A resulting in two proposals, namely, PSOHS#$A and APSOHS#$A for
AR !
6.:.2 E54)ri&)/$ ! R) (!$ / Di ( i#/
The PSO and APSO methodologies are both combined with S#$A for local search to mine ARs as described in previous section! The five datasets
used for all the other methodologies is adopted for generating ARs!
ARs are mined from the datasets using the two proposals and the PA of
the generated rules are plotted as shown in #igure =!B!
PSOHS#$A methodology of mining ARs performs better than simple
PSO in terms of PA for all the five datasets ta"en up for analysis! TheAPSOHS#$A methodology for mining ARs outperforms the other three
methods!
*&
*'
*(
*)
**
.&
.'
.(
.)
.*
+&&
PSO APSO PSO+SFLA APSO+SFLA
%atasets
Predictive Accurac!
8/9/2019 theis4
20/27
143
Fi"(r) 6.< Pr) i $i ) A (r y C#&4 ri #/ '#r ARM The fitness function ob4ective is ma%imization! /n order to enhance the
results the fitness values generated should be optimal! The fitness values for the proposed methodologies are plotted in #igure .!. for all the five datasets!
&
&/0
+
+/0
'
'/0
1
1/0
(
(/0
2enses %ataset
PSO APSO PSO+SFLA APSO+SFLA
Iteration 3um4er
Fitness 5alue
&
&/0
+
+/0
'
'/0
1
1/0
(
6a4erman7s Survival %ataset
PSO APSO PSO8SF2A APSO8SF2A
Iteration 3um4er
Fitness 5alue
8/9/2019 theis4
21/27
144
&&/0
++/0
''/0
11/0
((/0
0
9ar Evaluation %ataset
PSO APSO PSO8SF2A APSO8SF2A
Iteration 3um4er
Fitness 5alue
0
0.5
1
1.5
2
2.5
3
3.5
4
4.5
Postoperative Patient %ataset
PSO APSO PSO8SF2A APSO8SF2A
Iteration 3um4er
Fitness 5alue
&
&/0
+
+/0
'
'/0
1
1/0
(
:oo %ataset
PSO APSO PSO8SF2A APSO8SF2A
Iteration 3um4er
Fitness 5alue
8/9/2019 theis4
22/27
145
Fi"(r) 6.= Fi$/) !() '#r PSO 7i$% SFLA '#r ARM
The fitness values of both proposed methodologies* PSO HS#$A and
APSOHS#$A are more than the respective individual PSO and APSO values!
Thus, both methods perform better by generating ARs with enhanced PA, than
PSO and APSO methods!
The performances of the proposed two methods are compared with GA
and PSO methods discussed so far in terms of the PA of the ARs mined and the
results for the five datasets are shown in Table =!5!
The APSOHS#$A methodology outperforms the other methods for all
the five datasets by generating ARs with better PA! The APSO methodology
too generates ARs with optimal accuracy compared to other methods! The data
independent adaptation methodologies &SAPSO(, SAPSO5 and SA PSO+ ran"
ne%t in terms of performance for all the five datasets! 3owever the performance
of other methods varies among datasets considered in this study!
The number of rules generated by each methodology for the datasets
ta"en up for analysis is given in Table =! ! The SA PSO( methodology performs better among the data independent adaptation methodologies,
considered for analysis here as SAPSO!
The APSOHS#$A methodology of mining ARs generates more rules
than the other methods discussed! The SAPSO &SAPSO(+ performs better by
generating optimal number of ARs!
Thus the proposed APSOHS#$A methodology performs better whencompared to the other methods, in terms of PA and number of rules generated!
The S#$A performs effective local search thereby balancing between
e%ploration and e%ploitation and hence better performance!
T b!) 6.2 C#&4 ri #/ #' $%) Pr) i $i ) A (r y '#r ARM
GA AGAE!i$i $
GA PSO9PS
OCPSO NPSO SAPSO APSO GPS>
PSS
8/9/2019 theis4
23/27
145
.D )(!= )(!= . !DB )B!)( .B!D ) !( )B!)( ).!( .B!D )ation .B )@ )B )B!=( ))!) ))!.= )B!( ))!)5 ))!.5 )D!(5 ))
n
8/9/2019 theis4
24/27
146
6.6 SUMMARY
A hybrid method combining both genetic algorithm and Particle Swarm
Optimization called GPSO has been proposed! This method brings a balance between e%ploration and e%ploitation, resulting in higher prediction accuracy of
the ARs mined and consistency in performance! Two methodologies using PSO
with S#$A for local search &PSOHS#$A and APSOHS#$A+ has been proposed!
Among them, APSOHS#$A methodology generate ARs with better PA, than all
other methodologies discussed so far!
CHAPTER <CONCLUSIONS AND FUTURE 9OR@
8/9/2019 theis4
25/27
147
The salient conclusions of this research wor" and the scope for future wor" are
presented in this chapter!
!
&i+ Genetic Algorithm when used for mining ARs performs better than
other e%isting traditional methods!
&ii+ Particle swarm optimization when applied for mining ARs produce
results better than GA, but with minimum e%ecution time! The
increase in PA of PSO over GA is> B!.E for $enses dataset, B!@E for
3aberman
8/9/2019 theis4
26/27
148
&v+ GPSO methodology produces consistent results in comparison with
GA and PSO! The PA achieved is almost same throughout the
generations for each dataset, by utilizing GA to maintain diversity on
high ran"ed population and applying PSO
8/9/2019 theis4
27/27
149
e%tended to comple% and real life problems belonging to une%plored
application0domains, and the e%ecution time analysis of these methods
can be carried out! ethods to reduce the e%ecution time could be
e%plored!