Upload
harinima
View
220
Download
0
Embed Size (px)
Citation preview
8/21/2019 Dcmeet Third V1
1/81
1
Empirical Study on Mining Association Rules
Using Population Based Stochastic Search
Algorithms
RESEARCH PROPOSAL
K.INDIRA
Under the guidance of
Dr. S.KANMANIProfessor,
Department of Information Technology,
Pondicherry Engineering College
8/21/2019 Dcmeet Third V1
2/81
2
ORGANIZATION
OBJECTIVES
WHY ASSOCIATION RULES
INTRODUCTION
MOTIVATION
EMPRICAL STUDY
CONCLUSION
PUBLICATIONS
REFERNECES
8/21/2019 Dcmeet Third V1
3/81
3INDIAN INSTITUTE OFTECHNOLOGY, ROORKEE
1.OBJECTIVES
Exploration of usage of evolutionary algorithms for
mining association rules.
To develop efficient methodology for mining Association
rules both effectively and efficiently using population
based search methods namely Genetic Algorithm (GA) and
Particle Swarm Optimization(PSO).
To hybridize GA and PSO for mining association rules to
overcome each others weakness, leading to new
approach for association rule mining.
8/21/2019 Dcmeet Third V1
4/81
4
2.WHY ASSOCIATION RULE MINING
Large quantities of data is being accumulated.
There is a huge gap from the stored data to the
knowledge that could be construed from the data
Searching for meaningful information in largedatabases has become a very important issue.
Association rules, Clustering and Classification are
methods applied for extracting information fromdatabases.
Association rule mining is the most widely applied
method.
8/21/2019 Dcmeet Third V1
5/81
5
3. INTRODUCTION
3.1 DATA MINING
Extraction of interesting information orpatterns from data in large databases is known
as data mining.
8/21/2019 Dcmeet Third V1
6/81
6
3.2 ASSOCIATION RULE MINING
Association rule mining finds interesting
associations and/or correlation relationshipsamong large set of data items.
8/21/2019 Dcmeet Third V1
7/81
7
3.3 ASSOCIATION RULES
Association Rules are of form X Y with twocontrol parameters support and confidence
Support,s, probability that a transactioncontains X Y
Confidence, c,conditional probability that atransaction having X also contains Y
8/21/2019 Dcmeet Third V1
8/81
8
3.3 ASSOCIATION RULES
Tid Items bought
10 Milk, Nuts, Sugar
20 Milk, Coffee, Sugar
30 Milk, Sugar, Eggs
40 Nuts, Eggs, Bread50 Nuts, Coffee, Sugar , Eggs,
Bread
Customer
buys sugar
Customerbuys both
Customer
buys milk
Let minsup = 50%, minconf = 50%
Freq. Pat.: Milk:3, Nuts:3, Sugar:4, Eggs:3, {Milk, Sugar}:3
Association rules: Milk Sugar (60%, 100%) Sugar Milk (60%, 75%)
8/21/2019 Dcmeet Third V1
9/81
9
4. MOTIVATION
4.1 EXISTING SYSTEMApriori, FP Growth Tree, clat are some of the
popular algorithms for mining ARs.
Traverse the database many times.
I/O overhead, and computational complexity is
more.
Cannot meet the requirements of large-scale
database mining.
Does not fit in memory and is expensive to build
8/21/2019 Dcmeet Third V1
10/81
10
4.2 EVOLUTIONARY ALGORITHM
Applicable in problems where no (good) method is
available:
Discontinuities, non-linear constraints, multi-
modalities.
Discrete variable space.
Implicitly defined models (if-then-else constructs).
Noisy problems.
Most suitable in problems where multiple solutions are
required:
Multi-modal optimization problems.
Multi-objective optimization problems.
8/21/2019 Dcmeet Third V1
11/81
11
4.2 EVOLUTIONARY ALGORITHM
Parallel implementation is easier.
Evolutionary algorithms provide robust and efficient
approach in exploring large search space.
8/21/2019 Dcmeet Third V1
12/81
12
4.3 GA AND PSO : AN INTRODUCTION
Genetic algorithm (GA) and Particle swarm
optimization (PSO) are both population based
search methods and move from set of points
(population) to another set of points in a single
iteration with likely improvement using set of
control operators.
8/21/2019 Dcmeet Third V1
13/81
13
4.4 GENETIC ALGORITHM
A Genetic Algorithm (GA) is a procedure used to
find approximate solutions to search problems
through the application of the principles of
evolutionary biology.
8/21/2019 Dcmeet Third V1
14/81
14
4.2 PARTCILE SWARM OPTIMIZATION
PSOs mechanism is inspired by the social and
cooperative behavior displayed by various
species like birds, fish etc including human
beings.
8/21/2019 Dcmeet Third V1
15/81
Association Rule
(AR) Mining
Population BasedEvolutionary Methods
Genetic Algorithm
(GA)Particle Swarm
Optimization (PSO)
Mining Association
Rules using GA
Analyzing the roleofControl parameters in
GA for mining ARs
Mining ARs using
Self Adaptive GA
Elitist GA for
AssociationRule
Mining
Mining Association
rules with PSO
Mining Association
Rules with chaotic
PSO
Mining Association
rules with DynamicNeighborhood
Selection in PSO
Mining Associationrules with Self
Adaptive PSO
Hybrid GA/PSO
(GPSO) for AR
Mining
5. BLOCK DIAGRAM OF RESEARCH
MODULES
8/21/2019 Dcmeet Third V1
16/81
16
5.1 DATASETS DESCRIPTION
Lenses
Habermans Survival
Car Evaluation
Post operative care
Zoo
8/21/2019 Dcmeet Third V1
17/81
17
5.1 LENSES DATASET
Age of thepatient
1: Young 2: Presbyopic 3:Presbyopic
SpectaclePrescription
1: Myopic 2:Hypermetropic
Astigmatic 1: No 2: Yes
TearProduction
Rate
1: Reduced 2: Normal
Result 1: HardContactlenses
2: Soft ContactLenses
3: No lenses
8/21/2019 Dcmeet Third V1
18/81
18
5.1 HABERMANS SURVIVAL DATASET
Age of the patient 30-83Numeric
Patient's year ofoperation
NumericEg. 67
Number of positiveaxillary nodes
detected
0-46Numeric
Result 1= the patientsurvived 5 years or
longer s
2 = the patient diedwithin 5 year
8/21/2019 Dcmeet Third V1
19/81
19
5.1 CAR EVALUATION DATASET
Buying price Very high High Medium Low
MaintenancePrice
Very high High Medium Low
Doors 2 3 4 5
Persons 2 4 More
Luggage boot Small Big Medium
Safety Low Medium High
Result Unacceptable Acceptable Good Verygood
8/21/2019 Dcmeet Third V1
20/81
20
8/21/2019 Dcmeet Third V1
21/81
21
8/21/2019 Dcmeet Third V1
22/81
22
5.3.1 MINING AR USING GA
MethodologySelection : Tournament
Crossover Probability : Fixed ( Tested with 3 values)
Mutation Probability : No Mutation
Fitness Function :
Dataset : Lenses, Iris, Haberman from
UCI Irvinerepository.
Population : Fixed ( Tested with 3 values)
8/21/2019 Dcmeet Third V1
23/81
23
8/21/2019 Dcmeet Third V1
24/81
24
Flow chart of the GA
8/21/2019 Dcmeet Third V1
25/81
25
R lt A l i
8/21/2019 Dcmeet Third V1
26/81
Results Analysis
No. of Instances No. of Instances * 1.25 No. of Instances *1.5
Accuracy
%
No. of
Generations
Accuracy
%
No. of
Generations
Accuracy
%
No. of
Generations
Lenses 75 7 82 12 95 17
Haberman 71 114 68 88 64 70
Iris 77 88 87 53 82 45
Comparison based on variation in population Size.
Minimum Support & Minimum Confidence
Sup = 0.4 & con =0.4 Sup =0.9 & con =0.9 Sup = 0.9 & con = 0.2 Sup = 0.2 & con = 0.9
Accuracy
%
No. of
Gen
Accuracy
%
No. of
Gen.
Accuracy
%
No. of
Gen.
Accuracy
%
No. of
Gen
Lenses 22 20 49 11 70 21 95 18
Haberman 45 68 58 83 71 90 62 75
Iris 40 28 59 37 78 48 87 55
Comparison based on variation in Minimum Support and Confidence
8/21/2019 Dcmeet Third V1
27/81
27
8/21/2019 Dcmeet Third V1
28/81
28
Cross Over
Pc = .25 Pc = .5 Pc = .75
Accurac
y %
No. of
Generations
Accuracy % No. of
Generations
Accuracy
%
No. of
Generations
Lenses 95 8 95 16 95 13
Haberman 69 77 71 83 70 80
Iris 84 45 86 51 87 55
Dataset No. of
Instances
No. of
attributes
Population
Size
Minimum
Support
Minimum
confidence
Crossover
rate
Accuracy
in %
Lenses 24 4 36 0.2 0.9 0.25 95
Haberman 306 3 306 0.9 0.2 0.5 71
Iris 150 5 225 0.2 0.9 0.75 87
Comparison of the optimum value of Parameters for
maximum Accuracy achieved
Comparison based on variation in Crossover Probability
8/21/2019 Dcmeet Third V1
29/81
29
Population Size Vs Accuracy
Minimum Support and Confidence Vs Accuracy
8/21/2019 Dcmeet Third V1
30/81
30
Values of minimum support, minimum confidence and mutation
rate decides upon the accuracy of the system than other GA
parameters
Crossover rate affects the convergence rate rather than the
accuracy of the system
The optimum value of the GA parameters varies from data to
data and the fitness function plays a major role in optimizing the
results
Inferences
8/21/2019 Dcmeet Third V1
31/81
31
Mining ARs using Self Adaptive GA in Java.
Methodology
Selection : Roulette Wheel
Crossover Probability : Fixed ( Tested with 3 values)
Mutation Probability : Self Adaptive
Fitness Function :
Dataset : Lenses, Iris, Car from
UCI Irvine repository.
Population : Fixed ( Tested with 3 values)
Procedure SAGA
8/21/2019 Dcmeet Third V1
32/81
32
Procedure SAGA
Begin
Initialize population p(k);
Define the crossover and mutation rate;
Do
{
Do
{
Calculate support of all k rules;
Calculate confidence of all k rules;Obtain fitness;
Select individuals for crossover / mutation;
Calculate the average fitness of the n and (n-1) the generation;
Calculate the maximum fitness of the n and (n-1) the generation;
Based on the fitness of the selected item, calculate the new crossoverand mutation rate;
Choose the operation to be performed;
} k times;
}
8/21/2019 Dcmeet Third V1
33/81
Self Adaptive GA
SELF ADAPTIVE
Results Analysis
8/21/2019 Dcmeet Third V1
34/81
34
Dataset Traditional GA Self Adaptive GA
Accuracy No. of Generations Accuracy No. of Generations
Lenses 75 38 87.5 35
Haberman 52 36 68 28
Car Evaluation 85 29 96 21
Dataset Traditional GA Self Adaptive GA
Accuracy No. ofGenerations
Accuracy No. of Generations
Lenses 50 35 87.5 35
Haberman 36 38 68 28
Car
Evaluation
74 36 96 21
ACCURACY COMPARISON BETWEEN GA AND SAGA WHEN PARAMETERS ARE
SET TO TERMINATION OF SAGA
ACCURACY COMPARISON BETWEEN GA AND SAGA WHEN PARAMETERS ARE IDEAL
FOR TRADITIONAL GA
Results Analysis
ACCURACY COMPARISON BETWEEN GA AND SAGA WHEN
8/21/2019 Dcmeet Third V1
35/81
35
0
10
20
30
40
50
60
70
80
90
100
Lenses Haberman Car Evaluation
PredictiveAccuracy(%)
Dataset
Traditional GA Accuracy
Self Adaptive GA Accuracy
ACCURACY COMPARISON BETWEEN GA AND SAGA WHEN
PARAMETERS ARE IDEAL FOR TRADITIONAL GA
8/21/2019 Dcmeet Third V1
36/81
36
0
10
20
30
40
50
60
70
80
90
100
Lenses Haberman Car Evaluation
PredictiveAccuracy(%)
Dataset
Traditional GA
Self Adaptive GA
ACCURACY COMPARISON BETWEEN GA AND SAGA
WHEN PARAMETERS ARE ACCORDING TO
TERMINTAION OF SAGA
I f
8/21/2019 Dcmeet Third V1
37/81
Inferences
Self Adaptive GA gives better accuracy than Traditional GA.
37
8/21/2019 Dcmeet Third V1
38/81
GA with Elitism for Mining ARsMethodology
Selection : Elitism with roulette wheel
Crossover Probability : Fixed to Pc
Mutation Probability : Self Adaptive
Fitness Function : Fitness(x) = con(x)*(log(sup(x) *
length(x) + 1)
Dataset : Lenses, Iris, Car from UCI Irvine
repository.
Population : Fixed38
8/21/2019 Dcmeet Third V1
39/81
No. Of
Iterations
Lenses Car Evaluation Haberman
4 90 94.4 706 87.5 91.6 75
8 91.6 92.8 91.6
10 90 87.5 75
15 87.5 90 83.3
20 91.6 87.5 91.6
25 87.5 87.5 92.5
30 83.3 93.75 83.3
50 90 75 75
Predictive Accuracy for Mining AR based on GA with Elitism
Results Analysis
39
P di i A f Mi i AR b d GA
8/21/2019 Dcmeet Third V1
40/81
40
0
10
20
30
40
50
60
70
80
90
100
6 8 10 15 20 25 30 50
PredictiveAccuracy(%)
No. of Iterations
Lenses
Car Evaluation
Haberman
Predictive Accuracy for Mining AR based on GA
with Elitism
8/21/2019 Dcmeet Third V1
41/81
41
No of matches vs. No of iterations
8/21/2019 Dcmeet Third V1
42/81
No. Of
Iterations
Lenses (ms) Car Evaluation
(ms)
Haberman
(ms)
4 15 547 125
6 16 721 156
8 31 927 187
10 31 1104 20315 32 1525 281
20 47 1967 359
25 63 2504 421
30 78 2935 53050 94 4753 998
Execution Time for Mining AR based on GA with Elitism
42
8/21/2019 Dcmeet Third V1
43/81
43
0
1000
2000
3000
4000
5000
6000
4 6 8 10 15 20 25 30 50
Executiontime(ms)
No. of Iterations
Haberman (ms)
Car Evaluation (ms)
Lenses (ms)
8/21/2019 Dcmeet Third V1
44/81
Inferences
Marginally better accuracy arrived
Computational Efficiency found to be optimum
Elitism when introduced helps in retaining
chromosomes with good fitness values for next
generation
44
8/21/2019 Dcmeet Third V1
45/81
45
Mining ARs using PSO
Methodology
Each data itemset are represented as particles
The particles moves based on velocity
The particles position are updated based on
S O ( SO)
8/21/2019 Dcmeet Third V1
46/81
Particle Swarm Optimization (PSO)
Flow chart depicting the General PSO Algorithm:
Start
Initialize particles with random position
and velocity vectors.
For each particles position (p)evaluate fitness
If fitness(p) better than
fitness(pbest) then pbest= pLoopuntilall
particlesexhaus
t
Set best of pBests as gBest
Update particles velocity (eq. 1) and
position (eq. 3)
Loopun
tilmaxiter
Stop:giving gBest, optimal solution.
Results Analysis
8/21/2019 Dcmeet Third V1
47/81
Dataset NameTraditional
GA
Self
AdaptiveGA
PSO
Lenses 87.5 91.6 92.8
Haberman 75.5 92.5 91.6
Car evaluation 85 94.4 95
Results Analysis
0
200
400
600
800
1000
1200
4 6 8 10 15 20 25 30 50
Execution
Timemsec
No. of iterations
Haberman
PSO
SAGA0
20
40
60
80
100
4 6 8 101520253050
Executio
nTimemsec
No. of iterations
Lenses
PSO
SAGA
0
200
400
600
800
1000
1200
4 6 8 10 15 20 25 30 50
ExecutionTimemsec
No. of Iterations
Car Evaluation
PSO
SAGA
Predictive Accuracy
Execution Time
47
Inferences
8/21/2019 Dcmeet Third V1
48/81
Inferences
PSO produce results as effective as self adaptive GA
Computational effectiveness of PSO marginally fast when
compared to SAGA.
In PSO only the best particle passes information to others and
hence the computational capability of PSO is marginally better
than SAGA.
48BACK
Mining ARs using Chaotic PSO
8/21/2019 Dcmeet Third V1
49/81
49
Mining ARs using Chaotic PSO
The new chaotic map model is formulated as
Methodology
Initial point u0and V0to 0.1
The velocity of each particle is updated by
Mining ARs using
8/21/2019 Dcmeet Third V1
50/81
Compute xi(k+1)Compute (f(xi(k+1))
Reorder the particles
Generate neighborhoods I =1
k K
i = i +1
K = k+1
Start
K =1
Initialize xi(k), vi(k)
Compute f(xi(k))
Determine best particles in the
neighborhood of i
Update previous best if necessary
I N
Stop
no
no
yes
yes
Mining ARs using
Chaotic PSO
50
ACCURACY COMPARISON
8/21/2019 Dcmeet Third V1
51/81
51
80
82
84
86
88
90
92
94
96
98
100
Haberman lens car evaluation
PredictiveAccura
cy(%)
SAGA
PSO
CPSO
ACCURACY COMPARISON
C R t C i f L
8/21/2019 Dcmeet Third V1
52/81
52
75
80
85
90
95
100
SAGA pso cpso
4
6
8
10
15
20
25
30
50
Convergence Rate Comparison for Lenses
Convergence Rate Comparison for Car
8/21/2019 Dcmeet Third V1
53/81
53
40
50
60
70
80
90
100
SAGA pso cpso
4
6
8
10
15
20
25
30
50
g p
Evaluation
C R t C i f H b
8/21/2019 Dcmeet Third V1
54/81
54
0
10
20
30
40
50
60
70
80
90
100
SAGA pso cpso
4
6
8
10
15
20
25
30
50
Convergence Rate Comparison for Habermans
Survival
Inferences
8/21/2019 Dcmeet Third V1
55/81
Inferences
Better accuracy than PSO
The Chaotic Operators could be changed by altering the initial
values in chaotic operator function
The balance between exploration and exploitation is
maintained
55
Mining ARs using Neighborhood Selection
8/21/2019 Dcmeet Third V1
56/81
56
g g g
in PSOMethodology
The concept of local best particle (lbest) replacing the particle
best (pbest) is introduced
The neighborhood best (lbest) selection is as follows;
Calculate the distance of the current particle from other
particles
Find the nearest m particles as the neighbor of the current
particle based on distance calculated
Choose the local optimum lbest among the neighborhood
in terms of fitness values
8/21/2019 Dcmeet Third V1
57/81
57
Interestingness Measure
The interestingness measure for a rule is taken from relativeconfidence and is as follows:
Where k is the rule, x the antecedent part of the rule and y
the consequent part of the rule k.
Predictive Accuracy Comparison for Dynamic
8/21/2019 Dcmeet Third V1
58/81
58
88
89
90
91
92
93
94
95
96
97
98
Haberman lens car evaluation
PredictiveAcurac
y(%)
saga
pso
Npso
Neighborhood selection in PSO
8/21/2019 Dcmeet Third V1
59/81
59
Dataset Interestingness Value
Lens 0.82
Car Evaluation 0.73
HabermansSurvival 0.8
Measure of Interestingness
Execution Time Comparison for Dynamic
8/21/2019 Dcmeet Third V1
60/81
60
0
200
400
600
800
1000
1200
1400
1600
4 6 8 10 15 20 25 30 50
Lenses PSO
Lenses NPSO
Haberman's Survival PSO
Haberman's Survival NPSO
Car Evaluation PSO
Car Evaluation NPSO
Neighborhood selection in PSO
Predictive Accuracy over Generation for a) Car
8/21/2019 Dcmeet Third V1
61/81
61
0
10
20
30
40
50
60
70
80
90
100
PSO NPSO
PredictiveAccuracy(%)
4
6
8
10
15
20
25
30
50
Evaluation b) Lenses c) Habermans Survival datasets
0
10
20
30
40
50
60
70
80
90
100
PSO NPSO
PredictiveAccuracy(%)
4
6
8
10
15
20
25
30
50
0
10
20
30
40
50
60
70
80
90
100
PSO NPSO
PredictiveAccuracy(%)
4
6
8
10
15
20
25
30
50
Inferences
8/21/2019 Dcmeet Third V1
62/81
62
The avoidance of premature convergence at local optimalpoints tend to enhance the results
The selection of local best particles based on neighbors
(lbest) rather than particles own best (pbest) enhances
the accuracy of the rules mined
Inferences
Mining ARs using Self Adaptive
8/21/2019 Dcmeet Third V1
63/81
63
Chaotic PSO
A slight variant of the PSO is called inertia-weight PSO, in which a
weight parameter Is added to the velocity equation adopted
where, w is the inertia weight. The variable w plays the role of
balancing the global search and local search.A method of adaptive mutation rate is used
= m x
m x
min
) g/G
where, g is the generation index representing thecurrent number of evolutionary generations, and G is a
redefined maximum number of generations. Here, the
maximal and minimal weights max and min are usually
set to 0.9 and 0.4, respectively.
8/21/2019 Dcmeet Third V1
64/81
Effect of changing w
Dataset
Highest PA achieved within 50 runs of iterations
No weight
(Normal PSO)w = 0.5 w = 0.7
Lenses 87.5 88.09 84.75
Haberman 87.5 96.07 99.80
Car 96.4 99.88 99.84
POP Care 91.6 98.64 97.91
Zoo 83.3 96.88 98.97
8/21/2019 Dcmeet Third V1
65/81
Lenses
0
1020
30
40
50
60
70
80
90
100
5 10 15 25 50
PredictiveAccu
racy
No of generations
Predictive Accuracy CPSO
Predictive Accuracy
Weighted CPSO
Predictive Accuracy Self
Adaptive CPSO
8/21/2019 Dcmeet Third V1
66/81
Habermans Survival
75
80
85
90
95
100
5 10 15 25 50
PredictiveAccu
racy
No of generations
Predictive AccuracyCPSO
Predictive Accuracy
Weighted CPSO
Predictive Accuracy Self
Adaptive CPSO
8/21/2019 Dcmeet Third V1
67/81
Post operative Patient Care
0
1020
30
40
50
60
70
80
90
100
5 10 15 25 50
PredictiveAccuracy
No. of Generations
Predictive AccuracyCPSO
Predictive Accuracy
Weighted CPSO
Predictive Accuracy Self
Adaptive CPSO
8/21/2019 Dcmeet Third V1
68/81
Zoo
82
84
86
88
90
92
94
96
98
100
5 10 15 25 50
PredictiveAccu
racy
No. of Generations
Predictive AccuracyCPSO
Predictive Accuracy
Weighted CPSO
Predictive Accuracy Self
Adaptive CPSO
8/21/2019 Dcmeet Third V1
69/81
Car Evaluation
98.2
98.4
98.6
98.8
99
99.299.4
99.6
99.8
100
5 10 15 25 50
PredictiveAccu
racy
No of generations
Predictive Accuracy CPSO
Predictive Accuracy
Weighted CPSO
Predictive Accuracy Self
Adaptive CPSO
Inferences
8/21/2019 Dcmeet Third V1
70/81
70
In term of computational efficiency SACPSO is
faster than GA
Setting of appropriate values for the control
parameters involved in these heuristics methods
is the key point to success in these methods
Inferences
Mining AR using Hybrid GA/PSO
8/21/2019 Dcmeet Third V1
71/81
71
x
8/21/2019 Dcmeet Third V1
72/81
72
When Genetic algorithm used for mining association
rules Improvement in predictive accuracy achieved
Particle swarm optimization when adopted for mining
association rules produces results closer to GA but with
minimum execution time
The premature convergence being the major drawback
of PSO was handled by introducing inertia weights,chaotic maps, neighborhood selection adaptive inertia
weight
Papers Published
8/21/2019 Dcmeet Third V1
73/81
73
K.Indira, Dr.S.Kanmani, Framework for Comparison of Association Rule
Mining Using Genetic Algorithm, In : International Conference OnComputers, Communication & Intelligence , 2010.
K.Indira, Dr.S.Kanmani, Mining Association Rules Using Genetic Algorithm:
The role of Estimation Parameters , In : International conference on
advances in computing and communications, Communication in Computer
and Information Science, Springer LNCS,Volume 190, Part 8, 639-648, 2011
K.Indira, Dr. S. Kanmani , Gaurav Sethia.D, Kumaran.S, Prabhakar.J , Rule
Acquisition in Data Mining Using a Self Adaptive Genetic Algorithm, In :
First International conference on Computer Science and Information
Technology, Communication in Computer and Information Science, SpringerLNCS Volume 204, Part 1, 171-178, 2011.
K.Indira, Dr. S.Kanmani, Prasanth, Harish, Jeeva, Population Based Search
Methods in Mining Association Rules , In : Third International Conference
on Advances in Communication, Network, and Computing CNC 2012,
LNICST pp. 255261, 2012.
Conferences
J l
8/21/2019 Dcmeet Third V1
74/81
74
Journal
K.Indira, Dr. S.Kanmani, Performance Analysis of Genetic Algorithm for
Mining Association Rules, IJCSI International Journal of Computer Science
Issues, Vol. 9, Issue 2, No 1, 368-376, March 2012
K.Indira, Dr. S.Kanmani, Rule Acquisition using Genetic Algorithm,
accepted for publication in Journal of Computing
K.Indira, Dr. S.Kanmani, Enhancing Particle Swarm optimization using
chaotic operators for Association Rule Mining, communicated to
International Journal of Computer Science and Techniques
K.Indira, Dr. S.Kanmani, AssociationRule Mining by Dynamic Neighborhood
Selection in Particle Swarm Optimization, communicated to world science
Publications
References
8/21/2019 Dcmeet Third V1
75/81
Jing Li, Han Rui-feng, ASelf-Adaptive Genetic Algorithm Based On Real-
Coded, International Conference on Biomedical Engineering and
computer Science , Page(s): 1 - 4 , 2010
Chuan-Kang Ting, Wei-Ming Zeng, Tzu- Chieh Lin, LinkageDiscovery through
Data Mining, IEEE Magazine on Computational Intelligence, Volume 5,
February 2010.
Caises, Y., Leyva, E., Gonzalez, A., Perez, R., An extension of the Genetic
Iterative Approach for Learning Rule Subsets ,4th International Workshop
on Genetic and Evolutionary Fuzzy Systems, Page(s): 63 - 67 , 2010
Shangping Dai, Li Gao, Qiang Zhu, Changwu Zhu, A Novel Genetic Algorithm
Based on Image Databases for Mining Association Rules, 6th IEEE/ACIS
International Conference on Computer and Information Science, Page(s):977980, 2007
Peregrin, A., Rodriguez, M.A., Efficient Distributed Genetic Algorithm for
Rule Extraction,. Eighth International Conference on Hybrid Intelligent
Systems, HIS '08. Page(s): 531536, 2008 75
8/21/2019 Dcmeet Third V1
76/81
References Contd..
8/21/2019 Dcmeet Third V1
77/81
77
Maria J. Del Jesus, Jose A. Gamez, Pedro Gonzalez, Jose M. Puerta, On the
Discovery of Association Rules by means of Evolutionary Algorithms, from
Advanced Review of John Wiley & Sons , Inc. 2011
Junli Lu, Fan Yang, Momo Li, Lizhen Wang, Multi-objective Rule Discovery
Using the Improved Niched Pareto Genetic Algorithm, Third International
Conference on Measuring Technology and Mechatronics Automation, 2011.
Hamid Reza Qodmanan, Mahdi Nasiri, Behrouz Minaei-Bidgoli, MultiObjective Association Rule Mining with Genetic Algorithm without specifying
Minimum Support and Minimum Confidence, Expert Systems with
Applications 38 (2011) 288298.
Miguel Rodriguez, Diego M. Escalante, Antonio Peregrin, Efficient Distributed
Genetic Algorithm for Rule Extraction, Applied Soft Computing 11 (2011) 733743.
J.H. Ang, K.C. Tan , A.A. Mamun, An Evolutionary Memetic Algorithm for Rule
Extraction, Expert Systems with Applications 37 (2010) 13021315.
References Contd..
8/21/2019 Dcmeet Third V1
78/81
R.J. Kuo, C.M. Chao, Y.T. Chiu, Application of particle swarm optimization to
association rule mining, Applied Soft Computing 11 (2011) 326336.
Bilal Alatas , Erhan Akin, Multi-objective rule mining using a chaotic particle
swarm optimization algorithm, Knowledge-Based Systems 22 (2009) 455
460.
Mourad Ykhlef, A Quantum Swarm Evolutionary Algorithm for mining
association rules in large databases, Journal of King Saud University
Computer and Information Sciences (2011) 23, 16.
Haijun Su, Yupu Yang, Liang Zhao, Classification rule discovery with DE/QDE
algorithm, Expert Systems with Applications 37 (2010) 12161222.
Jing Li, Han Rui-feng, A Self-Adaptive Genetic Algorithm Based On Real-
Coded, International Conference on Biomedical Engineering and
computer Science , Page(s): 1 - 4 , 2010
Chuan-Kang Ting, Wei-Ming Zeng, Tzu- Chieh Lin, Linkage Discovery
through Data Mining, IEEE Magazine on Computational Intelligence,
Volume 5, February 2010.
78
References Contd..
8/21/2019 Dcmeet Third V1
79/81
79
Caises, Y., Leyva, E., Gonzalez, A., Perez, R., An extension of the Genetic
Iterative Approach for Learning Rule Subsets ,4th International Workshop
on Genetic and Evolutionary Fuzzy Systems, Page(s): 63 - 67 , 2010
Xiaoyuan Zhu, Yongquan Yu, Xueyan Guo, Genetic Algorithm Based on
Evolution Strategy and the Application in Data Mining, First International
Workshop on Education Technology and Computer Science, ETCS '09,
Volume: 1 , Page(s): 848852, 2009
Miguel Rodriguez, Diego M. Escalante, Antonio Peregrin, Efficient
Distributed Genetic Algorithm for Rule extraction, Applied Soft Computing
11 (2011) 733743.
Hamid Reza Qodmanan , Mahdi Nasiri, Behrouz Minaei-Bidgoli, Multi
objective association rule mining with genetic algorithm without specifyingminimum support and minimum confidence, Expert Systems with
Applications 38 (2011) 288298.
Yamina Mohamed Ben Ali, Soft Adaptive Particle Swarm Algorithm for Large
Scale Optimization, IEEE 2010.
References Contd..
8/21/2019 Dcmeet Third V1
80/81
80
Junli Lu, Fan Yang, Momo Li1, Lizhen Wang, Multi-objective Rule Discovery
Using the Improved Niched Pareto Genetic Algorithm, 2011 Third International
Conference on Measuring Technology and Mechatronics Automation.
Yan Chen, Shingo Mabu, Kotaro Hirasawa, Genetic relation algorithm with
guided mutation for the large-scale portfolio optimization, Expert Systems
with Applications 38 (2011), 33533363.
R.J. Kuo, C.M. Chao, Y.T. Chiu, Application of particle swarm optimization toassociation rule mining, Applied Soft Computing 11 (2011), 326336
Feng Lu, Yanfeng Ge, LiQun Gao, Self-adaptive Particle Swarm Optimization
Algorithm for Global Optimization, 2010 Sixth International Conference on
Natural Computation (ICNC 2010)
Fevrier Valdez, Patricia Melin, Oscar Castillo, An improved evolutionary
method with fuzzy logic for combining Particle Swarm Optimization and
Genetic Algorithms, Applied Soft Computing 11 (2011) ,26252632
8/21/2019 Dcmeet Third V1
81/81
hank You