Skip to main content
Log in

Association rule mining through adaptive parameter control in particle swarm optimization

  • Original Paper
  • Published:
Computational Statistics Aims and scope Submit manuscript

Abstract

Association rule mining is a data mining task on a great deal of academic research has been done and many algorithms are proposed. Association rule mining is treated as a twofold process by most of the methods. It increases the complexity of the system and takes up more time and space. Evolutionary Computation (EC) are fast growing search based optimization method for association rule mining. Among ECs particle swarm optimization (PSO) is more suited for mining association rules. The bottleneck of PSO is setting the precise values for their control parameters. Setting values to the control parameter is done either through parameter tuning or parameter control. This paper proposes an adaptive methodology for the control parameters in PSO namely, acceleration coefficients and inertia weight based on estimation of evolution state and fitness value respectively. Both of the proposed adaptive methods when tested on five datasets from University of California Irvine (UCI) repository proved to generate association rules with better accuracy and rule measures compared to simple PSO.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

References

  • Agrawal R, Imielinski T, Swami A (1993) Mining association rules between sets of items in large databases. In: Buneman P, Jajodia S (eds) Proceedings of the 1993 ACM SIGMOD international conference on management of data. ACM Press, New York, pp 207–216

    Chapter  Google Scholar 

  • Agrawal R, Srikant R (1994) Fast algorithms for mining association rules in large databases. In: Bocca JB, Jarke M, Zaniolo C (eds) Proceedings of the 20th international conference on very large data bases, Morgan Kaufmann, Burlington, pp 478–499

  • Back T (1992) Self-adaptation in genetic algorithms. In: Proceedings of the 1st European conference on artificial life, pp 263–271

  • Brin S, Motwani R, Ullman JD, Tsur S (1997) Dynamic itemset counting and implication rules for market basket data. In: Peckham J (eds) Proceedings of the 1997 ACM SIGMOD international conference on management of data, pp 255–264

  • Burke EK, Hyde M, Kendall G et al (2009) A survey of hyper-heuristics. In: Technical report NOTTCS-TR-SUB-0906241418-2747, School of computer science and information technology, University of Nottingham

  • Carlisle DG (2001) An off-the-shelf PSO. In: Proceeding on particle swarm optimization. Purdue School of Engineering and Technology, Indianapolis, pp 1–6

  • Ceglar A, Roddick JF (2006) Association mining. ACM Comput Surv 38(2):5

    Article  Google Scholar 

  • Chang CI, Chueh HE, Lin NP (2009) Sequential patterns mining with fuzzy time-intervals. In: Sixth international conference on fuzzy systems and knowledge discovery, pp 165–169

  • Chatterjee A, Siarry P (2006) Nonlinear inertia weight variation for dynamic adaptation in particle swarm optimization. Comput Oper Res 33:859–871

    Article  MATH  Google Scholar 

  • Clark P, Boswell R (1991) Rule induction with CN2: Some recent improvements. In: Proceedings of the European working session on learning, pp 151–163

  • Eberhart R, Kennedy J (1995) A new optimizer using particle swarm theory. In: Proceedings of the sixth international symposium on micro machine and human science. IEEE Service Center, Piscataway, NJ, pp 39–43

  • Eiben AE, Smith JE (2003) Introduction to evolutionary computing. Springer, Berlin

    Book  MATH  Google Scholar 

  • Eiben AE, Michalewicz Z, Schoenauer M, Smith JE (2007) Parameter control in evolutionary algorithms. In: Lobo FG, Lima CF, Michalewicz Z (eds) Parameter setting in evolutionary algorithms. Springer, Berlin, pp 19–46

    Chapter  Google Scholar 

  • Engelbrecht AP (2006) Fundamentals of computational swarm intelligence. Wiley, New York

    Google Scholar 

  • Good IJ (1965) The estimation of probabilities: an essay on modern Bayesian methods. MIT Press, Cambridge

  • Han J, Kamber M (2001) Data mining: concepts and techniques. Morgan Kaufmann, Burlington

    Google Scholar 

  • Han J, Pei J, Yin Y (2000) Mining frequent patterns without candidate generation, In: Proceeding of SIGMOD’00, Dallas, TX, May 2000, pp 1–12

  • Indira K, Kanmani S (2012) Association rule mining using self adaptive particle swarm optimization. IJCA Special Issue on Computational Intelligence and Information Security CIIS, pp 27–31

  • Indira K, Kanmani S et al (2012) Population Based Search Methods in Mining Association Rules. In: Third international conference on advances in communication, network, and computing—CNC 2012, LNCS, pp 255–261

  • Juang Y-T, Tung S-L, Chiu H-C (2011) Adaptive fuzzy particle swarm optimization for global optimization of multimodal functions. Inf Sci 181:4539–4549

    Article  MATH  MathSciNet  Google Scholar 

  • Kennedy J, Eberhart RC (2001) Swarm intelligence. Morgan Kaufman, Burlington

    Google Scholar 

  • Kennedy J, Eberhart RC (1995) Particle swarm optimization. In: Proceeding of the IEEE international conference on neural networks, pp 1942–1948

  • Kennedy J, Mendes R (2002) Population structure and particle swarm performance. In: Proceedings of the 2002 congress on evolutionary computation, pp 1671–1676

  • Kuo RJ, Chao CM, Chiu YT (2011) Application of particle swarm optimization to association rule mining. Appl Soft Comput 11(1):326–336

    Article  Google Scholar 

  • Leung SW, Yuen SY, Chow CK (2012) Parameter control system of evolutionary algorithm that is aided by the entire search history. Appl Soft Comput 12:3063–3078

    Article  Google Scholar 

  • Liu B, Abbas HA, McKay B (2003) Classification rule discovery with ant colony optimization. In: IEEE/WIC international conference on intelligent agent technology, 13–17 October, Halifax, Canada, pp 83–88

  • Liu B, Hsu W, Ma Y (1999) Pruning and summarizing the discovered associations. In: Proceeding of the fifth ACM SIGKDD international conference on Knowledge discovery and data mining, pp 125–136

  • Mahor A, Rangnekar S (2012) Short term generation scheduling of cascaded hydro electric system using novel self adaptive inertia weight PSO. Int J Electr Power Energy Syst 36(1):1–9

    Article  Google Scholar 

  • Mendes R, Kennedy J, Neves J (2004) The fully informed particle swarm: simpler, maybe better. IEEE Trans Evol Comput 8:204–210

    Article  Google Scholar 

  • Merz CJ, Murphy P (1996) UCI repository of machine learning databases. http://www.cs.uci.edu/mlearn/MLRepository.html

  • Nannen V, Smit SK, Eiben AE (2008) Costs and benefits of tuning parameters of evolutionary algorithms. In: Proceedings of the 10th international conference on parallel problem solving from nature. PPSN X, Dortmund, Germany, pp 528–538

  • Özbakır L, Baykasoglu A, Kulluk S, Yapıcı H (2010) TACO-miner: an ant colony based algorithm for rule extraction from trained neural networks. Expert Syst Appl 36(10):12295–12305

    Article  Google Scholar 

  • Özbakır L, Baykasoglu A, Kulluk S (2008) Rule extraction from neural networks via ant colony algorithm for data mining applications. In: Lecture notes in computer science 5313, pp 177–191

  • Parpinelli RS, Lopes HS, Freitas AA (2002) Data mining with an ant colony optimization algorithm. IEEE Trans Evol Comput 6(4):321–332 Special issue on Ant Colony Algorithms.

    Article  Google Scholar 

  • Parsopoulos KE, Vrahatis MN (2004) Unified particle swarm optimization scheme. In: Proceeding in lecture series on computational sciences, pp 868–873

  • Parsopoulos KE, Vrahatis MN (2007) Parameter selection and adaptation in unified particle swarm optimization. Math Comput Model 46:198–213

    Article  MATH  MathSciNet  Google Scholar 

  • Piatetsky-Shapiro G (1991) Discovery, analysis, and presentation of strong rules. In: Knowledge discovery in databases, AAAI/MIT Press, pp 229–248

  • Ratnaweera A, Halgamuge SK, Watson HC (2004) Self-organizing hierarchical particle swarm optimizer with time-varying acceleration coefficients. IEEE Trans Evol Comput 8(3):240–255

    Article  Google Scholar 

  • Ratnaweera A, Halgamuge S, Watson H (2003) Particle Swarm Optimization with self-adaptive acceleration coefficients. In: Proceedings of first international conference on fuzzy systems and knowledge discovery, pp 264–268

  • Senthil Arumugam M, Rao MVC (2008) On the improved performances of the particle swarm optimization algorithms with adaptive parameters, crossover operators and root mean square (RMS) variants for computing optimal control of a class of hybrid systems. Appl Soft Comput 8(1):324–336

  • Shenoy PD, Srinivasa KG, Venugopal KR, Patnaik LM (2003) Evolutionary approach for mining association rules on dynamic databases. In: Proceeding of Pacific-Asia conference on knowledge discovery and data mining, LNAI, 2637, Springer, Berlin, pp 325–336

  • Shi Y, Eberhart RC (2001) Fuzzy adaptive particle swarm optimization. Proc IEEE Congress Evolut Comput 1:101–106

    Google Scholar 

  • Shi Y, Eberhart RC (1998) Parameter selection in particle swarm optimization. In: Proceeding of the 7th conference on evolutionary programming, New York, pp 591–600

  • Shi Y, Eberhart RC (1999) Empirical study of particle swarm optimization. In: Proceedings of IEEE congress on evolutionary computation, pp 1945–1950

  • Song MP, Gu GC (2004) Research on particle swarm optimization: A review. In: Proceedings of the IEEE international conference on machine learning and cybernetics, pp 2236–2241

  • Tripathi PK, Bandyopadhyay KS, Pal SK (2007) Adaptive multi-objective particle swarm optimization agorithm. IN: Proceedings of IEEE congress on evolutionary computation, pp 2281–2288

  • Vrugt JA, Robinson BA, Hyman JM (2009) Self-adaptive multimethod search for global optimization in real-parameter spaces. IEEE Trans Evol Comput 13(2):243–259

    Article  Google Scholar 

  • Vzquez JM, lvarez Macas JL, Santos JCR (2002) Discovering numeric association rules via evolutionary algorithm. In: Pacific-Asia conference on knowledge discovery and data mining, pp 40–51

  • Wang Y, Li B, Weise T, Wang J, Yuan B, Tian Q (2011) Self-adaptive learning based particle swarm optimization. Inf Sci 181:4515–4538

    Article  MATH  Google Scholar 

  • Wang Y, Zhou J, Zhou C, Wang Y, Qin H, Youlin L (2012) An improved self-adaptive PSO technique for short-term hydrothermal scheduling. Expert Syst Appl 39(3):2288–2295

    Article  Google Scholar 

  • Xin J, Chen G, Hai Y (2009) A particle swarm optimizer with multistage linearly-decreasing inertia weight. In: International joint conference on computational sciences and optimization, pp 505–508

  • Yamaguchi T, Yasuda K (2006) Adaptive particle swarm optimization: self-coordinating mechanism with updating information. In: Proceedings of IEEE international conference on system, man, cybernatics, Taipei, Taiwan, pp 2303–2308

  • Zhan ZH, Xiao J, Zhang J, Chen WN (2007) Adaptive control of acceleration coefficients for particle swarm optimization based on clustering analysis. In: Proceedings of IEEE congress on evolutionary computation, Singapore, pp 3276–3282

  • Zhan Z-H, Zhang J, Li Y et al (2009) Adaptive particle swarm optimization. IEEE Trans Syst Man Cybern Part B Cybern 39(6):1362–1381

    Article  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to K. Indira.

Appendices

Appendix 1

1.1 UCI Datasets

The datasets that have been used in this research work have been taken from the UCI machine learning repository. The UCI Machine Learning Repository is a collection of databases, domain theories, and data generators that are used by the machine learning community for the empirical analysis of machine learning algorithms. The archive was created as an ftp archive in 1987 by David Aha and fellow graduate students at UC Irvine. The details of the datasets used in this study from http://archive.ics.uci.edu/ml/about.html are presented below.

Lenses dataset

  • Number of Instances: 24

  • Number of Attributes: 4 (all nominal)

  • Attribute Information: 3 Classes

  1. 1:

    the patient should be fitted with hard contact lenses,

  2. 2:

    the patient should be fitted with soft contact lenses,

  3. 3:

    the patient should not be fitted with contact lenses.

    • age of the patient: (1) young, (2) pre-presbyopic, (3) presbyopic

    • spectacle prescription: (1) myope, (2) hypermetrope

    • astigmatic: (1) no, (2) yes

    • tear production rate: (1) reduced, (2) normal

  • Number of Missing Attribute Values: 0

Haberman’s survival dataset

  • Number of Instances: 306

  • Number of Attributes: 4 (including the class attribute)

  • Attribute Information:

    1. 1.

      Age of patient at time of operation (numerical)

    2. 2.

      Patient’s year of operation (year—1900, numerical)

    3. 3.

      Number of positive axillary nodes detected (numerical)

    4. 4.

      Survival status (class attribute)

      • 1 = the patient survived 5 years or longer

      • 2 = the patient died within 5 year

  • Missing Attribute Values: None

Car evaluation dataset

  • Number of Instances: 1728 (instances completely cover the attribute space)

  • Number of Attributes: 6

  • Attribute Values:

    buying:

    v-high, high, med, low

    maint:

    v-high, high, med, low

    doors:

    2, 3, 4, 5-more

    persons:

    2, 4, more

    lug_boot:

    small, med, big

    safety:

    low, med, high

  • Missing Attribute Values: none

Post operative patient dataset

  • Number of Instances: 90

  • Number of Attributes: 9 including the decision (class attribute)

  • Attribute Information:

    1. 1.

      L-CORE (patient’s internal temperature in C):

    high (\(>\) 37), mid (\(>\)= 36 and \(<\)= 37), low (\(<\) 36)

    1. 2.

      L-SURF (patient’s surface temperature in C):

    high (\(>\) 36.5), mid (\(>\)= 36.5 and \(<\)= 35), low (\(<\) 35)

    1. 3.

      L-O2 (oxygen saturation in %):

    excellent (\(>\)= 98), good (\(>\)= 90 and \(<\) 98), fair (\(>\)= 80 and \(<\) 90), poor (\(<\) 80)

    1. 4.

      L-BP (last measurement of blood pressure):

    high (\(>\) 130/90), mid (\(<\)= 130/90 and \(>\)= 90/70), low (\(<\) 90/70)

    1. 5.

      SURF-STBL (stability of patient’s surface temperature):

    stable, mod-stable, unstable

    1. 6.

      CORE-STBL (stability of patient’s core temperature)

    stable, mod-stable, unstable

    1. 7.

      BP-STBL (stability of patient’s blood pressure)

    stable, mod-stable, unstable

    1. 8.

      COMFORT (patient’s perceived comfort at discharge, measured as

    an integer between 0 and 20)

    1. 9.

      decision ADM-DECS (discharge decision):

      • I (patient sent to Intensive Care Unit),

      • S (patient prepared to go home),

      • A (patient sent to general hospital floor)

  • Missing Attribute Values: Attribute 8 has 3 missing values

Zoo dataset

  • Number of Instances: 101

  • Number of Attributes: 18 (animal name, 15 Boolean attributes, 2 numerics)

  • Attribute Information: (name of attribute and type of value domain)

    1. animal name:

    Unique for each instance

    2. hair:

    Boolean

    3. feathers:

    Boolean

    4. eggs:

    Boolean

    5. milk:

    Boolean

    6. airborne:

    Boolean

    7. aquatic:

    Boolean

    8. predator:

    Boolean

    9. toothed:

    Boolean

    10. backbone:

    Boolean

    11. breathes:

    Boolean

    12. venomous:

    Boolean

    13. fins:

    Boolean

    14. legs:

    Numeric (set of values: {0,2,4,5,6,8})

    15. tail:

    Boolean

    16. domestic:

    Boolean

    17. catsize:

    Boolean

    18. type:

    Numeric (integer values in range [1,7])

  • Missing Attribute Values: None

Iris dataset

  • Number of Instances: 150 (50 in each of three classes)

  • Number of Attributes: 4 numeric, predictive attributes and the class

  • Attribute Information: Sepal length, Sepal width, Petal length and Petal width in cm

  • Class Types: Iris Setosa, Iris Versicolour and Iris Virginica.

  • Missing Attribute Values: None

Nursery dataset

  • Number of Instances: 12960 (instances completely cover the attribute space)

  • Number of Attributes: 8

  • Attribute Values:

    parents:

    usual, pretentious, great_pret

    has_nurs:

    proper, less_proper, improper, critical, very_crit

    form:

    complete, completed, incomplete, foster

    children:

    1, 2, 3, more

    housing:

    convenient, less_conv, critical

    finance:

    convenient, inconv

    social:

    non-prob, slightly_prob, problematic

    health:

    recommended, priority, not_recom

  • Missing Attribute Values: none

Tic Tac Toe dataset

  • Number of Instances: 958 (legal tic-tac-toe endgame boards)

  • Number of Attributes: 9, each corresponding to one tic-tac-toe square

  • Attribute Information: (x = player x has taken, o = player o has taken, b = blank)

    1. 1.

      top-left-square: {x,o,b}

    2. 2.

      top-middle-square: {x,o,b}

    3. 3.

      top-right-square: {x,o,b}

    4. 4.

      middle-left-square: {x,o,b}

    5. 5.

      middle-middle-square: {x,o,b}

    6. 6.

      middle-right-square: {x,o,b}

    7. 7.

      bottom-left-square: {x,o,b}

    8. 8.

      bottom-middle-square: {x,o,b}

    9. 9.

      bottom-right-square: {x,o,b}

    10. 10.

      Class: {positive,negative}

  • Missing Attribute Values: None

Wisconsin Breast Cancer dataset

  • Number of Instances: 699

  • Number of Attributes: 10 plus the class attribute

  • Attribute Information: Sample code number, Clump Thickness, Uniformity of Cell Size, Uniformity of Cell Shape, Marginal Adhesion, Single Epithelial Cell Size, Bare Nuclei, Bland Chromatin, Normal Nucleoli, Mitoses and Class (2 for benign, 4 for malignant).

  • Missing attribute values: 16

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Indira, K., Kanmani, S. Association rule mining through adaptive parameter control in particle swarm optimization. Comput Stat 30, 251–277 (2015). https://doi.org/10.1007/s00180-014-0533-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00180-014-0533-y

Keywords

Navigation