Dcmeet Third V1

Embed Size (px)

Citation preview

  • 8/21/2019 Dcmeet Third V1

    1/81

    1

    Empirical Study on Mining Association Rules

    Using Population Based Stochastic Search

    Algorithms

    RESEARCH PROPOSAL

    K.INDIRA

    Under the guidance of

    Dr. S.KANMANIProfessor,

    Department of Information Technology,

    Pondicherry Engineering College

  • 8/21/2019 Dcmeet Third V1

    2/81

    2

    ORGANIZATION

    OBJECTIVES

    WHY ASSOCIATION RULES

    INTRODUCTION

    MOTIVATION

    EMPRICAL STUDY

    CONCLUSION

    PUBLICATIONS

    REFERNECES

  • 8/21/2019 Dcmeet Third V1

    3/81

    3INDIAN INSTITUTE OFTECHNOLOGY, ROORKEE

    1.OBJECTIVES

    Exploration of usage of evolutionary algorithms for

    mining association rules.

    To develop efficient methodology for mining Association

    rules both effectively and efficiently using population

    based search methods namely Genetic Algorithm (GA) and

    Particle Swarm Optimization(PSO).

    To hybridize GA and PSO for mining association rules to

    overcome each others weakness, leading to new

    approach for association rule mining.

  • 8/21/2019 Dcmeet Third V1

    4/81

    4

    2.WHY ASSOCIATION RULE MINING

    Large quantities of data is being accumulated.

    There is a huge gap from the stored data to the

    knowledge that could be construed from the data

    Searching for meaningful information in largedatabases has become a very important issue.

    Association rules, Clustering and Classification are

    methods applied for extracting information fromdatabases.

    Association rule mining is the most widely applied

    method.

  • 8/21/2019 Dcmeet Third V1

    5/81

    5

    3. INTRODUCTION

    3.1 DATA MINING

    Extraction of interesting information orpatterns from data in large databases is known

    as data mining.

  • 8/21/2019 Dcmeet Third V1

    6/81

    6

    3.2 ASSOCIATION RULE MINING

    Association rule mining finds interesting

    associations and/or correlation relationshipsamong large set of data items.

  • 8/21/2019 Dcmeet Third V1

    7/81

    7

    3.3 ASSOCIATION RULES

    Association Rules are of form X Y with twocontrol parameters support and confidence

    Support,s, probability that a transactioncontains X Y

    Confidence, c,conditional probability that atransaction having X also contains Y

  • 8/21/2019 Dcmeet Third V1

    8/81

    8

    3.3 ASSOCIATION RULES

    Tid Items bought

    10 Milk, Nuts, Sugar

    20 Milk, Coffee, Sugar

    30 Milk, Sugar, Eggs

    40 Nuts, Eggs, Bread50 Nuts, Coffee, Sugar , Eggs,

    Bread

    Customer

    buys sugar

    Customerbuys both

    Customer

    buys milk

    Let minsup = 50%, minconf = 50%

    Freq. Pat.: Milk:3, Nuts:3, Sugar:4, Eggs:3, {Milk, Sugar}:3

    Association rules: Milk Sugar (60%, 100%) Sugar Milk (60%, 75%)

  • 8/21/2019 Dcmeet Third V1

    9/81

    9

    4. MOTIVATION

    4.1 EXISTING SYSTEMApriori, FP Growth Tree, clat are some of the

    popular algorithms for mining ARs.

    Traverse the database many times.

    I/O overhead, and computational complexity is

    more.

    Cannot meet the requirements of large-scale

    database mining.

    Does not fit in memory and is expensive to build

  • 8/21/2019 Dcmeet Third V1

    10/81

    10

    4.2 EVOLUTIONARY ALGORITHM

    Applicable in problems where no (good) method is

    available:

    Discontinuities, non-linear constraints, multi-

    modalities.

    Discrete variable space.

    Implicitly defined models (if-then-else constructs).

    Noisy problems.

    Most suitable in problems where multiple solutions are

    required:

    Multi-modal optimization problems.

    Multi-objective optimization problems.

  • 8/21/2019 Dcmeet Third V1

    11/81

    11

    4.2 EVOLUTIONARY ALGORITHM

    Parallel implementation is easier.

    Evolutionary algorithms provide robust and efficient

    approach in exploring large search space.

  • 8/21/2019 Dcmeet Third V1

    12/81

    12

    4.3 GA AND PSO : AN INTRODUCTION

    Genetic algorithm (GA) and Particle swarm

    optimization (PSO) are both population based

    search methods and move from set of points

    (population) to another set of points in a single

    iteration with likely improvement using set of

    control operators.

  • 8/21/2019 Dcmeet Third V1

    13/81

    13

    4.4 GENETIC ALGORITHM

    A Genetic Algorithm (GA) is a procedure used to

    find approximate solutions to search problems

    through the application of the principles of

    evolutionary biology.

  • 8/21/2019 Dcmeet Third V1

    14/81

    14

    4.2 PARTCILE SWARM OPTIMIZATION

    PSOs mechanism is inspired by the social and

    cooperative behavior displayed by various

    species like birds, fish etc including human

    beings.

  • 8/21/2019 Dcmeet Third V1

    15/81

    Association Rule

    (AR) Mining

    Population BasedEvolutionary Methods

    Genetic Algorithm

    (GA)Particle Swarm

    Optimization (PSO)

    Mining Association

    Rules using GA

    Analyzing the roleofControl parameters in

    GA for mining ARs

    Mining ARs using

    Self Adaptive GA

    Elitist GA for

    AssociationRule

    Mining

    Mining Association

    rules with PSO

    Mining Association

    Rules with chaotic

    PSO

    Mining Association

    rules with DynamicNeighborhood

    Selection in PSO

    Mining Associationrules with Self

    Adaptive PSO

    Hybrid GA/PSO

    (GPSO) for AR

    Mining

    5. BLOCK DIAGRAM OF RESEARCH

    MODULES

  • 8/21/2019 Dcmeet Third V1

    16/81

    16

    5.1 DATASETS DESCRIPTION

    Lenses

    Habermans Survival

    Car Evaluation

    Post operative care

    Zoo

  • 8/21/2019 Dcmeet Third V1

    17/81

    17

    5.1 LENSES DATASET

    Age of thepatient

    1: Young 2: Presbyopic 3:Presbyopic

    SpectaclePrescription

    1: Myopic 2:Hypermetropic

    Astigmatic 1: No 2: Yes

    TearProduction

    Rate

    1: Reduced 2: Normal

    Result 1: HardContactlenses

    2: Soft ContactLenses

    3: No lenses

  • 8/21/2019 Dcmeet Third V1

    18/81

    18

    5.1 HABERMANS SURVIVAL DATASET

    Age of the patient 30-83Numeric

    Patient's year ofoperation

    NumericEg. 67

    Number of positiveaxillary nodes

    detected

    0-46Numeric

    Result 1= the patientsurvived 5 years or

    longer s

    2 = the patient diedwithin 5 year

  • 8/21/2019 Dcmeet Third V1

    19/81

    19

    5.1 CAR EVALUATION DATASET

    Buying price Very high High Medium Low

    MaintenancePrice

    Very high High Medium Low

    Doors 2 3 4 5

    Persons 2 4 More

    Luggage boot Small Big Medium

    Safety Low Medium High

    Result Unacceptable Acceptable Good Verygood

  • 8/21/2019 Dcmeet Third V1

    20/81

    20

  • 8/21/2019 Dcmeet Third V1

    21/81

    21

  • 8/21/2019 Dcmeet Third V1

    22/81

    22

    5.3.1 MINING AR USING GA

    MethodologySelection : Tournament

    Crossover Probability : Fixed ( Tested with 3 values)

    Mutation Probability : No Mutation

    Fitness Function :

    Dataset : Lenses, Iris, Haberman from

    UCI Irvinerepository.

    Population : Fixed ( Tested with 3 values)

  • 8/21/2019 Dcmeet Third V1

    23/81

    23

  • 8/21/2019 Dcmeet Third V1

    24/81

    24

    Flow chart of the GA

  • 8/21/2019 Dcmeet Third V1

    25/81

    25

    R lt A l i

  • 8/21/2019 Dcmeet Third V1

    26/81

    Results Analysis

    No. of Instances No. of Instances * 1.25 No. of Instances *1.5

    Accuracy

    %

    No. of

    Generations

    Accuracy

    %

    No. of

    Generations

    Accuracy

    %

    No. of

    Generations

    Lenses 75 7 82 12 95 17

    Haberman 71 114 68 88 64 70

    Iris 77 88 87 53 82 45

    Comparison based on variation in population Size.

    Minimum Support & Minimum Confidence

    Sup = 0.4 & con =0.4 Sup =0.9 & con =0.9 Sup = 0.9 & con = 0.2 Sup = 0.2 & con = 0.9

    Accuracy

    %

    No. of

    Gen

    Accuracy

    %

    No. of

    Gen.

    Accuracy

    %

    No. of

    Gen.

    Accuracy

    %

    No. of

    Gen

    Lenses 22 20 49 11 70 21 95 18

    Haberman 45 68 58 83 71 90 62 75

    Iris 40 28 59 37 78 48 87 55

    Comparison based on variation in Minimum Support and Confidence

  • 8/21/2019 Dcmeet Third V1

    27/81

    27

  • 8/21/2019 Dcmeet Third V1

    28/81

    28

    Cross Over

    Pc = .25 Pc = .5 Pc = .75

    Accurac

    y %

    No. of

    Generations

    Accuracy % No. of

    Generations

    Accuracy

    %

    No. of

    Generations

    Lenses 95 8 95 16 95 13

    Haberman 69 77 71 83 70 80

    Iris 84 45 86 51 87 55

    Dataset No. of

    Instances

    No. of

    attributes

    Population

    Size

    Minimum

    Support

    Minimum

    confidence

    Crossover

    rate

    Accuracy

    in %

    Lenses 24 4 36 0.2 0.9 0.25 95

    Haberman 306 3 306 0.9 0.2 0.5 71

    Iris 150 5 225 0.2 0.9 0.75 87

    Comparison of the optimum value of Parameters for

    maximum Accuracy achieved

    Comparison based on variation in Crossover Probability

  • 8/21/2019 Dcmeet Third V1

    29/81

    29

    Population Size Vs Accuracy

    Minimum Support and Confidence Vs Accuracy

  • 8/21/2019 Dcmeet Third V1

    30/81

    30

    Values of minimum support, minimum confidence and mutation

    rate decides upon the accuracy of the system than other GA

    parameters

    Crossover rate affects the convergence rate rather than the

    accuracy of the system

    The optimum value of the GA parameters varies from data to

    data and the fitness function plays a major role in optimizing the

    results

    Inferences

  • 8/21/2019 Dcmeet Third V1

    31/81

    31

    Mining ARs using Self Adaptive GA in Java.

    Methodology

    Selection : Roulette Wheel

    Crossover Probability : Fixed ( Tested with 3 values)

    Mutation Probability : Self Adaptive

    Fitness Function :

    Dataset : Lenses, Iris, Car from

    UCI Irvine repository.

    Population : Fixed ( Tested with 3 values)

    Procedure SAGA

  • 8/21/2019 Dcmeet Third V1

    32/81

    32

    Procedure SAGA

    Begin

    Initialize population p(k);

    Define the crossover and mutation rate;

    Do

    {

    Do

    {

    Calculate support of all k rules;

    Calculate confidence of all k rules;Obtain fitness;

    Select individuals for crossover / mutation;

    Calculate the average fitness of the n and (n-1) the generation;

    Calculate the maximum fitness of the n and (n-1) the generation;

    Based on the fitness of the selected item, calculate the new crossoverand mutation rate;

    Choose the operation to be performed;

    } k times;

    }

  • 8/21/2019 Dcmeet Third V1

    33/81

    Self Adaptive GA

    SELF ADAPTIVE

    Results Analysis

  • 8/21/2019 Dcmeet Third V1

    34/81

    34

    Dataset Traditional GA Self Adaptive GA

    Accuracy No. of Generations Accuracy No. of Generations

    Lenses 75 38 87.5 35

    Haberman 52 36 68 28

    Car Evaluation 85 29 96 21

    Dataset Traditional GA Self Adaptive GA

    Accuracy No. ofGenerations

    Accuracy No. of Generations

    Lenses 50 35 87.5 35

    Haberman 36 38 68 28

    Car

    Evaluation

    74 36 96 21

    ACCURACY COMPARISON BETWEEN GA AND SAGA WHEN PARAMETERS ARE

    SET TO TERMINATION OF SAGA

    ACCURACY COMPARISON BETWEEN GA AND SAGA WHEN PARAMETERS ARE IDEAL

    FOR TRADITIONAL GA

    Results Analysis

    ACCURACY COMPARISON BETWEEN GA AND SAGA WHEN

  • 8/21/2019 Dcmeet Third V1

    35/81

    35

    0

    10

    20

    30

    40

    50

    60

    70

    80

    90

    100

    Lenses Haberman Car Evaluation

    PredictiveAccuracy(%)

    Dataset

    Traditional GA Accuracy

    Self Adaptive GA Accuracy

    ACCURACY COMPARISON BETWEEN GA AND SAGA WHEN

    PARAMETERS ARE IDEAL FOR TRADITIONAL GA

  • 8/21/2019 Dcmeet Third V1

    36/81

    36

    0

    10

    20

    30

    40

    50

    60

    70

    80

    90

    100

    Lenses Haberman Car Evaluation

    PredictiveAccuracy(%)

    Dataset

    Traditional GA

    Self Adaptive GA

    ACCURACY COMPARISON BETWEEN GA AND SAGA

    WHEN PARAMETERS ARE ACCORDING TO

    TERMINTAION OF SAGA

    I f

  • 8/21/2019 Dcmeet Third V1

    37/81

    Inferences

    Self Adaptive GA gives better accuracy than Traditional GA.

    37

  • 8/21/2019 Dcmeet Third V1

    38/81

    GA with Elitism for Mining ARsMethodology

    Selection : Elitism with roulette wheel

    Crossover Probability : Fixed to Pc

    Mutation Probability : Self Adaptive

    Fitness Function : Fitness(x) = con(x)*(log(sup(x) *

    length(x) + 1)

    Dataset : Lenses, Iris, Car from UCI Irvine

    repository.

    Population : Fixed38

  • 8/21/2019 Dcmeet Third V1

    39/81

    No. Of

    Iterations

    Lenses Car Evaluation Haberman

    4 90 94.4 706 87.5 91.6 75

    8 91.6 92.8 91.6

    10 90 87.5 75

    15 87.5 90 83.3

    20 91.6 87.5 91.6

    25 87.5 87.5 92.5

    30 83.3 93.75 83.3

    50 90 75 75

    Predictive Accuracy for Mining AR based on GA with Elitism

    Results Analysis

    39

    P di i A f Mi i AR b d GA

  • 8/21/2019 Dcmeet Third V1

    40/81

    40

    0

    10

    20

    30

    40

    50

    60

    70

    80

    90

    100

    6 8 10 15 20 25 30 50

    PredictiveAccuracy(%)

    No. of Iterations

    Lenses

    Car Evaluation

    Haberman

    Predictive Accuracy for Mining AR based on GA

    with Elitism

  • 8/21/2019 Dcmeet Third V1

    41/81

    41

    No of matches vs. No of iterations

  • 8/21/2019 Dcmeet Third V1

    42/81

    No. Of

    Iterations

    Lenses (ms) Car Evaluation

    (ms)

    Haberman

    (ms)

    4 15 547 125

    6 16 721 156

    8 31 927 187

    10 31 1104 20315 32 1525 281

    20 47 1967 359

    25 63 2504 421

    30 78 2935 53050 94 4753 998

    Execution Time for Mining AR based on GA with Elitism

    42

  • 8/21/2019 Dcmeet Third V1

    43/81

    43

    0

    1000

    2000

    3000

    4000

    5000

    6000

    4 6 8 10 15 20 25 30 50

    Executiontime(ms)

    No. of Iterations

    Haberman (ms)

    Car Evaluation (ms)

    Lenses (ms)

  • 8/21/2019 Dcmeet Third V1

    44/81

    Inferences

    Marginally better accuracy arrived

    Computational Efficiency found to be optimum

    Elitism when introduced helps in retaining

    chromosomes with good fitness values for next

    generation

    44

  • 8/21/2019 Dcmeet Third V1

    45/81

    45

    Mining ARs using PSO

    Methodology

    Each data itemset are represented as particles

    The particles moves based on velocity

    The particles position are updated based on

    S O ( SO)

  • 8/21/2019 Dcmeet Third V1

    46/81

    Particle Swarm Optimization (PSO)

    Flow chart depicting the General PSO Algorithm:

    Start

    Initialize particles with random position

    and velocity vectors.

    For each particles position (p)evaluate fitness

    If fitness(p) better than

    fitness(pbest) then pbest= pLoopuntilall

    particlesexhaus

    t

    Set best of pBests as gBest

    Update particles velocity (eq. 1) and

    position (eq. 3)

    Loopun

    tilmaxiter

    Stop:giving gBest, optimal solution.

    Results Analysis

  • 8/21/2019 Dcmeet Third V1

    47/81

    Dataset NameTraditional

    GA

    Self

    AdaptiveGA

    PSO

    Lenses 87.5 91.6 92.8

    Haberman 75.5 92.5 91.6

    Car evaluation 85 94.4 95

    Results Analysis

    0

    200

    400

    600

    800

    1000

    1200

    4 6 8 10 15 20 25 30 50

    Execution

    Timemsec

    No. of iterations

    Haberman

    PSO

    SAGA0

    20

    40

    60

    80

    100

    4 6 8 101520253050

    Executio

    nTimemsec

    No. of iterations

    Lenses

    PSO

    SAGA

    0

    200

    400

    600

    800

    1000

    1200

    4 6 8 10 15 20 25 30 50

    ExecutionTimemsec

    No. of Iterations

    Car Evaluation

    PSO

    SAGA

    Predictive Accuracy

    Execution Time

    47

    Inferences

  • 8/21/2019 Dcmeet Third V1

    48/81

    Inferences

    PSO produce results as effective as self adaptive GA

    Computational effectiveness of PSO marginally fast when

    compared to SAGA.

    In PSO only the best particle passes information to others and

    hence the computational capability of PSO is marginally better

    than SAGA.

    48BACK

    Mining ARs using Chaotic PSO

  • 8/21/2019 Dcmeet Third V1

    49/81

    49

    Mining ARs using Chaotic PSO

    The new chaotic map model is formulated as

    Methodology

    Initial point u0and V0to 0.1

    The velocity of each particle is updated by

    Mining ARs using

  • 8/21/2019 Dcmeet Third V1

    50/81

    Compute xi(k+1)Compute (f(xi(k+1))

    Reorder the particles

    Generate neighborhoods I =1

    k K

    i = i +1

    K = k+1

    Start

    K =1

    Initialize xi(k), vi(k)

    Compute f(xi(k))

    Determine best particles in the

    neighborhood of i

    Update previous best if necessary

    I N

    Stop

    no

    no

    yes

    yes

    Mining ARs using

    Chaotic PSO

    50

    ACCURACY COMPARISON

  • 8/21/2019 Dcmeet Third V1

    51/81

    51

    80

    82

    84

    86

    88

    90

    92

    94

    96

    98

    100

    Haberman lens car evaluation

    PredictiveAccura

    cy(%)

    SAGA

    PSO

    CPSO

    ACCURACY COMPARISON

    C R t C i f L

  • 8/21/2019 Dcmeet Third V1

    52/81

    52

    75

    80

    85

    90

    95

    100

    SAGA pso cpso

    4

    6

    8

    10

    15

    20

    25

    30

    50

    Convergence Rate Comparison for Lenses

    Convergence Rate Comparison for Car

  • 8/21/2019 Dcmeet Third V1

    53/81

    53

    40

    50

    60

    70

    80

    90

    100

    SAGA pso cpso

    4

    6

    8

    10

    15

    20

    25

    30

    50

    g p

    Evaluation

    C R t C i f H b

  • 8/21/2019 Dcmeet Third V1

    54/81

    54

    0

    10

    20

    30

    40

    50

    60

    70

    80

    90

    100

    SAGA pso cpso

    4

    6

    8

    10

    15

    20

    25

    30

    50

    Convergence Rate Comparison for Habermans

    Survival

    Inferences

  • 8/21/2019 Dcmeet Third V1

    55/81

    Inferences

    Better accuracy than PSO

    The Chaotic Operators could be changed by altering the initial

    values in chaotic operator function

    The balance between exploration and exploitation is

    maintained

    55

    Mining ARs using Neighborhood Selection

  • 8/21/2019 Dcmeet Third V1

    56/81

    56

    g g g

    in PSOMethodology

    The concept of local best particle (lbest) replacing the particle

    best (pbest) is introduced

    The neighborhood best (lbest) selection is as follows;

    Calculate the distance of the current particle from other

    particles

    Find the nearest m particles as the neighbor of the current

    particle based on distance calculated

    Choose the local optimum lbest among the neighborhood

    in terms of fitness values

  • 8/21/2019 Dcmeet Third V1

    57/81

    57

    Interestingness Measure

    The interestingness measure for a rule is taken from relativeconfidence and is as follows:

    Where k is the rule, x the antecedent part of the rule and y

    the consequent part of the rule k.

    Predictive Accuracy Comparison for Dynamic

  • 8/21/2019 Dcmeet Third V1

    58/81

    58

    88

    89

    90

    91

    92

    93

    94

    95

    96

    97

    98

    Haberman lens car evaluation

    PredictiveAcurac

    y(%)

    saga

    pso

    Npso

    Neighborhood selection in PSO

  • 8/21/2019 Dcmeet Third V1

    59/81

    59

    Dataset Interestingness Value

    Lens 0.82

    Car Evaluation 0.73

    HabermansSurvival 0.8

    Measure of Interestingness

    Execution Time Comparison for Dynamic

  • 8/21/2019 Dcmeet Third V1

    60/81

    60

    0

    200

    400

    600

    800

    1000

    1200

    1400

    1600

    4 6 8 10 15 20 25 30 50

    Lenses PSO

    Lenses NPSO

    Haberman's Survival PSO

    Haberman's Survival NPSO

    Car Evaluation PSO

    Car Evaluation NPSO

    Neighborhood selection in PSO

    Predictive Accuracy over Generation for a) Car

  • 8/21/2019 Dcmeet Third V1

    61/81

    61

    0

    10

    20

    30

    40

    50

    60

    70

    80

    90

    100

    PSO NPSO

    PredictiveAccuracy(%)

    4

    6

    8

    10

    15

    20

    25

    30

    50

    Evaluation b) Lenses c) Habermans Survival datasets

    0

    10

    20

    30

    40

    50

    60

    70

    80

    90

    100

    PSO NPSO

    PredictiveAccuracy(%)

    4

    6

    8

    10

    15

    20

    25

    30

    50

    0

    10

    20

    30

    40

    50

    60

    70

    80

    90

    100

    PSO NPSO

    PredictiveAccuracy(%)

    4

    6

    8

    10

    15

    20

    25

    30

    50

    Inferences

  • 8/21/2019 Dcmeet Third V1

    62/81

    62

    The avoidance of premature convergence at local optimalpoints tend to enhance the results

    The selection of local best particles based on neighbors

    (lbest) rather than particles own best (pbest) enhances

    the accuracy of the rules mined

    Inferences

    Mining ARs using Self Adaptive

  • 8/21/2019 Dcmeet Third V1

    63/81

    63

    Chaotic PSO

    A slight variant of the PSO is called inertia-weight PSO, in which a

    weight parameter Is added to the velocity equation adopted

    where, w is the inertia weight. The variable w plays the role of

    balancing the global search and local search.A method of adaptive mutation rate is used

    = m x

    m x

    min

    ) g/G

    where, g is the generation index representing thecurrent number of evolutionary generations, and G is a

    redefined maximum number of generations. Here, the

    maximal and minimal weights max and min are usually

    set to 0.9 and 0.4, respectively.

  • 8/21/2019 Dcmeet Third V1

    64/81

    Effect of changing w

    Dataset

    Highest PA achieved within 50 runs of iterations

    No weight

    (Normal PSO)w = 0.5 w = 0.7

    Lenses 87.5 88.09 84.75

    Haberman 87.5 96.07 99.80

    Car 96.4 99.88 99.84

    POP Care 91.6 98.64 97.91

    Zoo 83.3 96.88 98.97

  • 8/21/2019 Dcmeet Third V1

    65/81

    Lenses

    0

    1020

    30

    40

    50

    60

    70

    80

    90

    100

    5 10 15 25 50

    PredictiveAccu

    racy

    No of generations

    Predictive Accuracy CPSO

    Predictive Accuracy

    Weighted CPSO

    Predictive Accuracy Self

    Adaptive CPSO

  • 8/21/2019 Dcmeet Third V1

    66/81

    Habermans Survival

    75

    80

    85

    90

    95

    100

    5 10 15 25 50

    PredictiveAccu

    racy

    No of generations

    Predictive AccuracyCPSO

    Predictive Accuracy

    Weighted CPSO

    Predictive Accuracy Self

    Adaptive CPSO

  • 8/21/2019 Dcmeet Third V1

    67/81

    Post operative Patient Care

    0

    1020

    30

    40

    50

    60

    70

    80

    90

    100

    5 10 15 25 50

    PredictiveAccuracy

    No. of Generations

    Predictive AccuracyCPSO

    Predictive Accuracy

    Weighted CPSO

    Predictive Accuracy Self

    Adaptive CPSO

  • 8/21/2019 Dcmeet Third V1

    68/81

    Zoo

    82

    84

    86

    88

    90

    92

    94

    96

    98

    100

    5 10 15 25 50

    PredictiveAccu

    racy

    No. of Generations

    Predictive AccuracyCPSO

    Predictive Accuracy

    Weighted CPSO

    Predictive Accuracy Self

    Adaptive CPSO

  • 8/21/2019 Dcmeet Third V1

    69/81

    Car Evaluation

    98.2

    98.4

    98.6

    98.8

    99

    99.299.4

    99.6

    99.8

    100

    5 10 15 25 50

    PredictiveAccu

    racy

    No of generations

    Predictive Accuracy CPSO

    Predictive Accuracy

    Weighted CPSO

    Predictive Accuracy Self

    Adaptive CPSO

    Inferences

  • 8/21/2019 Dcmeet Third V1

    70/81

    70

    In term of computational efficiency SACPSO is

    faster than GA

    Setting of appropriate values for the control

    parameters involved in these heuristics methods

    is the key point to success in these methods

    Inferences

    Mining AR using Hybrid GA/PSO

  • 8/21/2019 Dcmeet Third V1

    71/81

    71

    x

  • 8/21/2019 Dcmeet Third V1

    72/81

    72

    When Genetic algorithm used for mining association

    rules Improvement in predictive accuracy achieved

    Particle swarm optimization when adopted for mining

    association rules produces results closer to GA but with

    minimum execution time

    The premature convergence being the major drawback

    of PSO was handled by introducing inertia weights,chaotic maps, neighborhood selection adaptive inertia

    weight

    Papers Published

  • 8/21/2019 Dcmeet Third V1

    73/81

    73

    K.Indira, Dr.S.Kanmani, Framework for Comparison of Association Rule

    Mining Using Genetic Algorithm, In : International Conference OnComputers, Communication & Intelligence , 2010.

    K.Indira, Dr.S.Kanmani, Mining Association Rules Using Genetic Algorithm:

    The role of Estimation Parameters , In : International conference on

    advances in computing and communications, Communication in Computer

    and Information Science, Springer LNCS,Volume 190, Part 8, 639-648, 2011

    K.Indira, Dr. S. Kanmani , Gaurav Sethia.D, Kumaran.S, Prabhakar.J , Rule

    Acquisition in Data Mining Using a Self Adaptive Genetic Algorithm, In :

    First International conference on Computer Science and Information

    Technology, Communication in Computer and Information Science, SpringerLNCS Volume 204, Part 1, 171-178, 2011.

    K.Indira, Dr. S.Kanmani, Prasanth, Harish, Jeeva, Population Based Search

    Methods in Mining Association Rules , In : Third International Conference

    on Advances in Communication, Network, and Computing CNC 2012,

    LNICST pp. 255261, 2012.

    Conferences

    J l

  • 8/21/2019 Dcmeet Third V1

    74/81

    74

    Journal

    K.Indira, Dr. S.Kanmani, Performance Analysis of Genetic Algorithm for

    Mining Association Rules, IJCSI International Journal of Computer Science

    Issues, Vol. 9, Issue 2, No 1, 368-376, March 2012

    K.Indira, Dr. S.Kanmani, Rule Acquisition using Genetic Algorithm,

    accepted for publication in Journal of Computing

    K.Indira, Dr. S.Kanmani, Enhancing Particle Swarm optimization using

    chaotic operators for Association Rule Mining, communicated to

    International Journal of Computer Science and Techniques

    K.Indira, Dr. S.Kanmani, AssociationRule Mining by Dynamic Neighborhood

    Selection in Particle Swarm Optimization, communicated to world science

    Publications

    References

  • 8/21/2019 Dcmeet Third V1

    75/81

    Jing Li, Han Rui-feng, ASelf-Adaptive Genetic Algorithm Based On Real-

    Coded, International Conference on Biomedical Engineering and

    computer Science , Page(s): 1 - 4 , 2010

    Chuan-Kang Ting, Wei-Ming Zeng, Tzu- Chieh Lin, LinkageDiscovery through

    Data Mining, IEEE Magazine on Computational Intelligence, Volume 5,

    February 2010.

    Caises, Y., Leyva, E., Gonzalez, A., Perez, R., An extension of the Genetic

    Iterative Approach for Learning Rule Subsets ,4th International Workshop

    on Genetic and Evolutionary Fuzzy Systems, Page(s): 63 - 67 , 2010

    Shangping Dai, Li Gao, Qiang Zhu, Changwu Zhu, A Novel Genetic Algorithm

    Based on Image Databases for Mining Association Rules, 6th IEEE/ACIS

    International Conference on Computer and Information Science, Page(s):977980, 2007

    Peregrin, A., Rodriguez, M.A., Efficient Distributed Genetic Algorithm for

    Rule Extraction,. Eighth International Conference on Hybrid Intelligent

    Systems, HIS '08. Page(s): 531536, 2008 75

  • 8/21/2019 Dcmeet Third V1

    76/81

    References Contd..

  • 8/21/2019 Dcmeet Third V1

    77/81

    77

    Maria J. Del Jesus, Jose A. Gamez, Pedro Gonzalez, Jose M. Puerta, On the

    Discovery of Association Rules by means of Evolutionary Algorithms, from

    Advanced Review of John Wiley & Sons , Inc. 2011

    Junli Lu, Fan Yang, Momo Li, Lizhen Wang, Multi-objective Rule Discovery

    Using the Improved Niched Pareto Genetic Algorithm, Third International

    Conference on Measuring Technology and Mechatronics Automation, 2011.

    Hamid Reza Qodmanan, Mahdi Nasiri, Behrouz Minaei-Bidgoli, MultiObjective Association Rule Mining with Genetic Algorithm without specifying

    Minimum Support and Minimum Confidence, Expert Systems with

    Applications 38 (2011) 288298.

    Miguel Rodriguez, Diego M. Escalante, Antonio Peregrin, Efficient Distributed

    Genetic Algorithm for Rule Extraction, Applied Soft Computing 11 (2011) 733743.

    J.H. Ang, K.C. Tan , A.A. Mamun, An Evolutionary Memetic Algorithm for Rule

    Extraction, Expert Systems with Applications 37 (2010) 13021315.

    References Contd..

  • 8/21/2019 Dcmeet Third V1

    78/81

    R.J. Kuo, C.M. Chao, Y.T. Chiu, Application of particle swarm optimization to

    association rule mining, Applied Soft Computing 11 (2011) 326336.

    Bilal Alatas , Erhan Akin, Multi-objective rule mining using a chaotic particle

    swarm optimization algorithm, Knowledge-Based Systems 22 (2009) 455

    460.

    Mourad Ykhlef, A Quantum Swarm Evolutionary Algorithm for mining

    association rules in large databases, Journal of King Saud University

    Computer and Information Sciences (2011) 23, 16.

    Haijun Su, Yupu Yang, Liang Zhao, Classification rule discovery with DE/QDE

    algorithm, Expert Systems with Applications 37 (2010) 12161222.

    Jing Li, Han Rui-feng, A Self-Adaptive Genetic Algorithm Based On Real-

    Coded, International Conference on Biomedical Engineering and

    computer Science , Page(s): 1 - 4 , 2010

    Chuan-Kang Ting, Wei-Ming Zeng, Tzu- Chieh Lin, Linkage Discovery

    through Data Mining, IEEE Magazine on Computational Intelligence,

    Volume 5, February 2010.

    78

    References Contd..

  • 8/21/2019 Dcmeet Third V1

    79/81

    79

    Caises, Y., Leyva, E., Gonzalez, A., Perez, R., An extension of the Genetic

    Iterative Approach for Learning Rule Subsets ,4th International Workshop

    on Genetic and Evolutionary Fuzzy Systems, Page(s): 63 - 67 , 2010

    Xiaoyuan Zhu, Yongquan Yu, Xueyan Guo, Genetic Algorithm Based on

    Evolution Strategy and the Application in Data Mining, First International

    Workshop on Education Technology and Computer Science, ETCS '09,

    Volume: 1 , Page(s): 848852, 2009

    Miguel Rodriguez, Diego M. Escalante, Antonio Peregrin, Efficient

    Distributed Genetic Algorithm for Rule extraction, Applied Soft Computing

    11 (2011) 733743.

    Hamid Reza Qodmanan , Mahdi Nasiri, Behrouz Minaei-Bidgoli, Multi

    objective association rule mining with genetic algorithm without specifyingminimum support and minimum confidence, Expert Systems with

    Applications 38 (2011) 288298.

    Yamina Mohamed Ben Ali, Soft Adaptive Particle Swarm Algorithm for Large

    Scale Optimization, IEEE 2010.

    References Contd..

  • 8/21/2019 Dcmeet Third V1

    80/81

    80

    Junli Lu, Fan Yang, Momo Li1, Lizhen Wang, Multi-objective Rule Discovery

    Using the Improved Niched Pareto Genetic Algorithm, 2011 Third International

    Conference on Measuring Technology and Mechatronics Automation.

    Yan Chen, Shingo Mabu, Kotaro Hirasawa, Genetic relation algorithm with

    guided mutation for the large-scale portfolio optimization, Expert Systems

    with Applications 38 (2011), 33533363.

    R.J. Kuo, C.M. Chao, Y.T. Chiu, Application of particle swarm optimization toassociation rule mining, Applied Soft Computing 11 (2011), 326336

    Feng Lu, Yanfeng Ge, LiQun Gao, Self-adaptive Particle Swarm Optimization

    Algorithm for Global Optimization, 2010 Sixth International Conference on

    Natural Computation (ICNC 2010)

    Fevrier Valdez, Patricia Melin, Oscar Castillo, An improved evolutionary

    method with fuzzy logic for combining Particle Swarm Optimization and

    Genetic Algorithms, Applied Soft Computing 11 (2011) ,26252632

  • 8/21/2019 Dcmeet Third V1

    81/81

    hank You