Elsevier

Information Sciences

Volume 288, 20 December 2014, Pages 254-278
Information Sciences

Combining human analysis and machine data mining to obtain credible data relations

https://doi.org/10.1016/j.ins.2014.08.014Get rights and content

Abstract

Can a model constructed using data mining (DM) programs be trusted? It is known that a decision-tree model can contain relations that are statistically significant, but, in reality, meaningless to a human. When the task is domain analysis, meaningless relations are problematic, since they can lead to wrong conclusions and can consequently undermine a human’s trust in DM programs. To eliminate problematic relations from the conclusions of analysis, we propose an interactive method called Human–Machine Data Mining (HMDM). The method constructs multiple models in a specific way so that a human can reexamine the relations in different contexts and, based on observed evidence, conclude which relations and models are credible—that is, both meaningful and of high quality. Based on the extracted credible relations and models, the human can construct correct overall conclusions about the domain. The method is demonstrated in two complex domains, extracting credible relations and models that indicate the segments of the higher education sector and the research and development sector that influence the economic welfare of a country. An experimental evaluation shows that the method is capable of finding important relations and models that are better in both meaning and quality than those constructed solely by the DM programs.

Introduction

In data mining (DM) and machine learning (ML), a human supplies the data and tunes the parameters of used methods. The obtained model is typically the result of several iterative, parameter-tuning steps. This paper aims to improve the interaction between humans and DM and ML programs and, therefore, belongs to the field of interactive DM (IDM) or interactive ML (IML) (the terms are used interchangeably in the literature) [71], [73].

The goal of IML is to “help scientists and engineers exploit more of their specialized data” [53]. IML “focuses on methods that empower domain experts to control and direct machine learning tools from within the deployed environment, whereas traditional machine learning does this in the development environment” [53].

The field of IML has recently received a great deal of attention. The preface of the IUI 2013 Workshop on Interactive Machine Learning stated, “Many applications of Machine Learning (ML) involve interactions with humans… a growing community of researchers at the intersection of ML and human–computer interaction are making interaction with humans a central part of developing ML systems. These efforts include applying interaction design principles to ML systems, using human-subject testing to evaluate ML systems and inspire new methods, and changing the input and output channels of ML systems to better leverage human capabilities” [54]. The mission statement of one of the Microsoft research groups dealing with IML notes: “… with the advancement of computational techniques such as machine learning, we now have the unprecedented ability to embed ‘smarts’ that allow machines to assist users in completing their tasks. We believe that trying to fully automate tasks is extremely difficult and even undesirable, but instead there exists a computational design methodology which allows us to gracefully combine automated services with direct user manipulation” [12].

When supervised DM and ML methods construct models in complex domains, such as economic and social domains, the models often contain less-credible relations [1], [27], [57]. Here, the relation is a pattern that connects a set of attributes describing the properties of a concept underlying the data with a class/target attribute, which represents the concept. The term less-credible means that the relation is of either low quality or high quality, but is meaningless to a human analyst. Meaningless means that a relation’s semantic is contradictory to the human’s common sense or domain knowledge, and a meaningless state can only be determined by including the human in the DM process. When the task is domain analysis, less-credible relations must be eliminated from the constructed models, since they lead to incorrect conclusions about the most important relations in the domain and can, consequently, undermine the human’s trust in the DM system [59].

The problem is illustrated by the example in Fig. 1. The decision-tree model on the right side of Fig. 1 represents a domain model. The tree is constructed from the data (the table on the left side of Fig. 1) using the J48 algorithm in Weka [69] with default parameters. The first three columns, or attributes, of this table represent properties of a person, while the final column, or class, indicates the person’s gender. Each row, or example, represents a person. In the tree, the node represents an attribute, and the leaves represent the class. In each leaf, the number in brackets represents the number of examples that reach that leaf. The tree contains a single relation, indicating that a person is a woman if the person has long hair and that a person is a man if the person has short hair. The relation is of high quality, since the tree’s accuracy (ACC) is 100%. ACC denotes the overall performance of the tree, expressed as the percentage of correctly classified examples out of all the examples classified by the tree. The relation is meaningless, however, since several men have long hair but are not women (as the left branch of the tree suggests).

The problem that this paper examines most commonly stems from an incompleteness of data [52]. For example, adding more rows and columns to the table in Fig. 1 would likely result in a different relation, but adding the right additional data might be a demanding task. Humans, however, can detect weak relations in domain models using domain knowledge and common sense.

The knowledge that men and women have long and short hair is objective in terms of common sense, as is the case in Fig. 1, but it is hard to take a purely objective position when humans are involved. Humans can also be subjective in terms of fairness; however this discussion is beyond the scope of this paper.

Although the relation in Fig. 1 is of high quality, its meaninglessness makes it less credible.

Another example was obtained through DM in a real-life domain. The decision-tree model presented in Fig. 2 is constructed with the J48 algorithm in Weka using the default parameters and a minimum number of instances per leaf (MNIL) of 5. The tree is constructed from a data set composed of 37 attributes describing the research and development (R&D) sector of a country, 167 examples representing countries and the class that differentiates countries according to their economic welfare into “low”, “middle” and “high” (see Section 4.1 for more information on this data set). In the tree, the subtrees form the relations. In each leaf, the first number in brackets represents the number of examples that reach that leaf. The second number represents the number of the examples of the class value other than the one represented by the leaf. The quantities are expressed in decimals to account for the weights of the examples with missing values.

The tree contains three interesting relations. The first is that countries with better welfare invest extensively in R&D. The relation contains attribute “GERD per capita (PPP$)” (GERD stands for Gross Domestic Expenditure on R&D and PPP$ for purchasing power parity in American dollars), which represents the level of investment in R&D. This relation appears twice in the tree. Both times, the “higher than” side of the subtree (>10.8 and >105.5) leads to leaves representing welfare better than that on the “less than” side. One could conclude that the first relation is a valid candidate for a credible relation in the tree because it is meaningful; that is, it is in accordance with domain knowledge [63] and common sense, it appears twice in the tree and, both times, it makes a clear distinction between countries with different levels of welfare. This relation is marked in bold in Fig. 2. The second relation—“Sector investing the most in R&D” (the right subtree)—seems to be meaningless, since all but one of the leaves represent the class “high”, and the single “middle” leaf represents the countries for which the sector is unknown (“N/A” value). Therefore, the entire subtree can be replaced with a single node: “high”. A detailed analysis shows that the problem is caused by several missing values, resulting in a relation that is statistically correct but meaningless to humans. The third relation—“Sector employing the most researchers”—distinguishes between “low” and “middle” countries. However, for “middle” countries, any sector could be the main employer, which makes the relation meaningless. Due to their lack of meaning, the second and third relations are not shown in bold. However, these relations should be verified with additional tests.

To eliminate less-credible relations from the models, both automatic and interactive approaches were suggested. Examples of the former include the pruning of decision trees [56], maximum-ambiguity-based sample selection for tree construction [67], alternative-node-selection measures in trees [70], fuzzy-entropy-maximization-based classification rule refinement [66], and the correction of a quality estimate to eliminate the random rules with optimistically high values of quality [47]. Typical examples of the latter suggest improvements in the form of new training examples [18] or a list of attributes that better describe the class [59]. The presented approaches aim to improve the model’s predictive performance by allowing meaningless relations to remain a part of the model, as long as they positively influence the quality. The resulting models are not acceptable when the task is domain analysis.

In contrast to the presented approaches, we propose a method that constructs multiple models (for example, decision trees) in an algorithmic manner, in order to examine the models’ relations for credibility. In this process, a human observes relations in different contexts and, based on common sense, informal knowledge about the domain, the observed relations’ frequency, and the stability and quality of the models in which the relations appear, concludes which relations are credible and which are not. Credible relations are then extracted through a specifically designed relation-extraction scheme for overall conclusions. In parallel, credible models, composed of credible relations, are also extracted.

Our method, which we have named Human–Machine Data Mining (HMDM), combines DM and human knowledge with the main motivation of transforming the rather ad-hoc process of DM into a systematic procedure. The primary purpose of HMDM is a domain analysis that increases the human understanding of the domain; therefore, HMDM aims to improve the process of finding, not just the models with the best predictive performance, but credible relations and models. The information that the human gathers from the process is just as valuable as the information gained from the results.

The benefits of using HMDM are as follows:

  • Humans contribute their knowledge from the real world compensating for incomplete data.

  • Computers contribute their immense computing speed and generate new hypotheses under human supervision.

  • Humans contribute by directing computer searches into interesting hypothesis spaces.

  • Humans perform in-depth analysis on existing data and design superior models in the real world.

The contributions of this paper are: (1) a novel IML method (HMDM) for extracting credible relations and models from data, based on an interactive and iterative process that exploits the advantages of humans and machine algorithms; (2) an extension of the corrected class probability estimate statistical measure, originally conceived for classification rules, that allows it to work on decision trees; (3) interactive explanations of DM results, conceived to facilitate the extraction of credible relations and models; (4) a demonstration of the HMDM method on two real-life domains; (5) an evaluation of the interactive method through a user study.

The remaining part of the paper is structured as follows. The relevant literature is reviewed in Section 2. The interactive method (HMDM) for the extraction of credible relations and models is defined in Section 3. Section 4 demonstrates the method in the two complex macroeconomic domains. The method is evaluated in Section 5. Section 6 concludes the paper and proposes future work.

Section snippets

Literature review

In recent years, several methods have been designed to improve the cooperation between humans and DM methods in interactive car systems [48], video retrieval [38], e-mail categorization [17], [59], object recognition [18], sensory recommendations [13], and citizen-science projects [31]. The literature indicates that involving a human in DM can significantly improve the obtained results. For example, Kapoor et al. [29] proposed an interactive method for the iterative readjustment of a kernel

Human–machine data mining

According to constructivist learning theory, the best learning framework for a human is experimental task-based learning, which enables the human to actively integrate the acquired knowledge into his/her existing mental model of the domain or to use the acquired knowledge to challenge and modify his/her existing model. Whereas a single model constructed with classic DM is comparable to a teacher delivering a didactic lecture covering the subject matter, the IDM method that we propose is

Real-life examples

This section presents two applications of the HMDM method on the real-life domains of higher education and R&D. First, the domains and their corresponding data sets are introduced. Second, the experimental setup is described. Third, the step-by-step example of applying HMDM to the higher education domain is presented, followed by the conclusions drawn from the analysis. Finally, we present the conclusions from the analysis of the R&D domain.

Evaluation

The procedures in HMDM that are based on attribute selection mechanisms are similar to those in WAS, since they all construct multiple attribute sets and models to obtain the best results. Therefore, we begin the evaluation by drawing a line between the two methods. We then present an experimental setup, results of comparisons and a discussion of the results. This is followed by a user study, which was conducted to understand whether humans accept credible models as better than automatically

Conclusions and discussion

In this paper, we present a new method: Human–Machine Data Mining (HMDM). Its primary advantage is based on the interaction between the two most advanced information mechanisms: the brute force of computers enriched with DM, and human insight and understanding. The implemented interactive system constructs multiple models and arranges them in deleted and attached attributes graphs. The human observes the constructed models and, with the help of the relation-extraction scheme, extracts credible

Acknowledgements

The authors would like to thank the anonymous reviewers for their valuable comments and suggestions to improve the quality of the paper. We are also grateful to Sanja Kovač, who contributed to the experimental evaluation, and Jure Grabnar, who contributed to the HMDM program.

Vedrana Vidulin is a researcher at the Department of Intelligent Systems at Jožef Stefan Institute, Ljubljana, Slovenia. She received her PhD degree at the Jožef Stefan International Postgraduate School, Ljubljana, Slovenia. Her research interests are related to data mining and machine learning, human–computer interaction, ambient intelligence and automatic web genre identification. Her major scientific achievement is the Human–Machine Data Mining method, which she has applied to several

References (74)

  • D. Martens et al.

    Performance of classification models from a user perspective

    Decis. Support Syst.

    (2011)
  • F. Nasoz et al.

    Affectively intelligent and adaptive car interfaces

    Inform. Sci.

    (2010)
  • K.M. Osei-Bryson

    Evaluation of decision trees: a multi-criteria approach

    Comput. Oper. Res.

    (2004)
  • S. Stumpf et al.

    Interacting meaningfully with machine learning systems: three experiments

    Int. J. Hum.–Comput. St.

    (2009)
  • E. Štrumbelj et al.

    Explaining instance classifications with interactions of subsets of feature values

    Data Knowl. Eng.

    (2009)
  • N.C. Varsakelis

    Education, political institutions and innovative activity: a cross-country empirical investigation

    Res. Policy

    (2006)
  • M. Ware et al.

    Interactive machine learning: letting users build classifiers

    Int. J. Hum.–Comput. St.

    (2001)
  • S. Badr et al.

    Case study of inaccuracies in the granulation of decision trees

    Soft. Comput.

    (2011)
  • B. Becker et al.

    Visualizing the simple bayesian classifier

  • J. Black et al.

    A Dictionary of Economics

    (2009)
  • M. Bohanec et al.

    Trading accuracy for simplicity in decision trees

    Mach. Learn.

    (1994)
  • L. Breiman

    Random forests

    Mach. Learn.

    (2001)
  • L. Cao et al.

    Domain Driven Data Mining

    (2010)
  • J. Cohen

    A coefficient of agreement for nominal scales

    Educ. Psychol. Meas.

    (1960)
  • D.A. Cohn et al.

    Active learning with statistical models

    J. Artif. Intell. Res.

    (1996)
  • Computational User Experiences Microsoft Research Group, Interactive Machine Learning,...
  • E. Costello, L. McGinty, Supporting user interaction in flavor sampling trials, in: Proc. of Workshop on Intelligence...
  • M.W. Craven, Extracting Comprehensible Models from Trained Neural Networks, PhD thesis, School of Computer Science,...
  • J. Demšar, B. Zupan, G. Leban, T. Curk, Orange: from experimental machine learning to interactive data mining, in:...
  • K. Dey, S. Rosenthal, M. Veloso, Using interaction to improve intelligence: how intelligent systems should ask users...
  • J.A. Fails, D.R. Olsen Jr., Interactive machine learning, in: Proc. of the 8th Int. Conf. on Intelligent User...
  • A.J. Feelders

    Prior knowledge in economic applications of data mining

  • I. Guyon et al.

    An introduction to variable and feature selection

    J. Mach. Learn. Res.

    (2003)
  • Y. Huang, T.M. Mitchell, Text clustering with extended user feedback, in: Proc. of the Int. SIGIR Conf. on R&D in...
  • A. Jakulin, Machine Learning Based on Attribute Interaction, PhD thesis, Faculty of Computer and Information Science,...
  • D. Jensen et al.

    Multiple comparisons in induction algorithms

    Mach. Learn.

    (2000)
  • G.H. John, Enhancements to the Data Mining Process, PhD thesis, Computer Science Department, Stanford University, CA,...
  • Cited by (0)

    Vedrana Vidulin is a researcher at the Department of Intelligent Systems at Jožef Stefan Institute, Ljubljana, Slovenia. She received her PhD degree at the Jožef Stefan International Postgraduate School, Ljubljana, Slovenia. Her research interests are related to data mining and machine learning, human–computer interaction, ambient intelligence and automatic web genre identification. Her major scientific achievement is the Human–Machine Data Mining method, which she has applied to several domains in the fields of macroeconomic research, automatic web genre identification and demography.

    Marko Bohanec is a senior researcher at the Department of Knowledge Technologies, Jožef Stefan Institute, Ljubljana, Slovenia, and a professor of computer science at the University of Nova Gorica, Slovenia. His major research interests are related to decision support systems, data mining and, particularly, multi-criteria modeling, machine learning and the integration of data mining and decision support. He has developed a number of decision support tools and systems, such as DECMAK, DEX and DEXi, and applied them in management, economy, ecology, agronomy, medicine and healthcare.

    Matjaž Gams is the head of Department of Intelligent Systems at Jožef Stefan Institute and a professor of computer science at the University of Ljubljana and Jožef Stefan International Postgraduate School (IPS), Slovenia. He received his degrees at the University of Ljubljana and IPS. He teaches or has taught on 10 faculties in Slovenia and Germany. His professional interests include intelligent systems, artificial intelligence, cognitive science, intelligent agents, business intelligence and information society. He is a member of the editorial boards of 11 journals and is the managing director of the journal Informatica. He is also a co-founder of various societies in Slovenia, such as the Engineering Academy, AI Society and Cognitive Society, and is the president or secretary of various societies, including ACM Slovenia. His major scientific achievement is the discovery of the principle of multiple knowledge.

    View full text