data mining and knowledge discovery: a review of …data mining and knowledge discovery: a review of...

40
Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach Ryszard S. Michalski and Kenneth A. Kaufman P97-3 MLI 97-2 March 1997

Upload: others

Post on 20-Aug-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

Data Mining and Knowledge Discovery:

A Review of Issues and

a Multistrategy Approach

Ryszard S. Michalski and Kenneth A. Kaufman

P97-3

MLI 97-2

March 1997

Page 2: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

Data Mining and Knowledge Discovery:A Review of Issues and a Multistrategy Approach∗

Ryszard S. Michalski and Kenneth A. Kaufman

Abstract

An enormous proliferation of databases in almost every area of human endeavor has createda great demand for new, powerful tools for turning data into useful, task-orientedknowledge. In efforts to satisfy this need, researchers have been exploring ideas andmethods developed in machine learning, pattern recognition, statistical data analysis, datavisualization, neural nets, etc. These efforts have led to the emergence of a new researcharea, frequently called data mining and knowledge discovery. The first part of this paper isa compendium of ideas on the applicability of symbolic machine learning methods to thisarea. The second part describes a multistrategy methodology for conceptual dataexploration, by which we mean the derivation of high-level concepts and descriptions fromdata through symbolic reasoning involving both data and background knowledge. Themethodology, which has been implemented in the INLEN system, combines machinelearning, database and knowledge-based technologies. To illustrate the system’scapabilities, we present results from its application to a problem of discovery of economicand demographic patterns in a database containing facts and statistics about the countries ofthe world. The presented results demonstrate a high potential utility of the methodology forassisting a user in solving practical data mining and knowledge discovery tasks.

Keywords: Multistrategy Learning, Integrated Learning, Knowledge Discovery,Conceptual Clustering, Inductive Generalization. Constructive Induction, TemporalReasoning, Structured Data

Acknowledgments

This research was done in the Machine Learning and Inference Laboratory of GeorgeMason University. The authors thank their past and current collaborators from theLaboratory, who provided research feedback, comments, and assistance at different stagesof the development of many of the ideas and computer programs described in this chapter.In particular, we thank Eric Bloedorn, Mark Maloof, Jim Ribeiro, Janusz Wnek, and QiZhang for their participation in these research efforts. Support for the Machine Learningand Inference Laboratory’s research activities has been provided in part by the NationalScience Foundation under Grants No. DMI-9496192 and IRI-9020266, in part by theOffice of Naval Research under Grant No. N00014-91-J-1351, in part by the DefenseAdvanced Research Projects Agency under Grant No. N00014-91-J-1854 administered bythe Office of Naval Research, and in part by the Defense Advanced Research ProjectsAgency under Grants No. F49620-92-J-0549 and F49620-95-1-0462 administered by theAir Force Office of Scientific Research. ∗ This paper is based on the authors’ Chapter 2 of Machine Learning and Data Mining: Methods andApplications, edited by R.S. Michalski, I. Bratko and M. Kubat, John Wiley & Sons publishers, 1997 (toappear).

Page 3: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 Introduction

The current information age is characterized by an extraordinary expansion of data that arebeing generated and stored about all kinds of human endeavors. An increasing proportionof these data is recorded in the form of computer databases, in order that the computertechnology may easily access it. The availability of very large volumes of such data hascreated a problem of how to extract from them useful, task-oriented knowledge.

Data analysis techniques that have been traditionally used for such tasks include regressionanalysis, cluster analysis, numerical taxonomy, multidimensional analysis, othermultivariate statistical methods, stochastic models, time series analysis, nonlinearestimation techniques, and others (e.g., Daniel and Wood, 1980; Tukey, 1986;Morganthaler and Tukey, 1989; Diday, 1989; Sharma, 1996). These techniques have beenwidely used for solving many practical problems. They are, however, primarily orientedtoward the extraction of quantitative and statistical data characteristics, and as such haveinherent limitations.

For example, a statistical analysis can determine covariances and correlations betweenvariables in data. It cannot, however, characterize the dependencies at an abstract,conceptual level, and produce a causal explanation of reasons why these dependenciesexist. Nor can it develop a justification of these relationships in the form of higher-levellogic-style descriptions and laws. A statistical data analysis can determine the centraltendency and variance of given factors, and a regression analysis can fit a curve to a set ofdatapoints. These techniques cannot, however, produce a qualitative description of theregularities and determine their dependence on factors not explicitly provided in the data,nor can they draw an analogy between the discovered regularity and a regularity in anotherdomain.

A numerical taxonomy technique can create a classification of entities, and specify anumerical similarity among the entities assembled into the same or different categories. Itcannot, however, build qualitative descriptions of the classes created and hypothesizereasons for the entities being in the same category. Attributes that define the similarity, aswell as the similarity measures, must be defined by a data analyst in advance. Also, thesetechniques cannot by themselves draw upon background domain knowledge in order toautomatically generate relevant attributes and determine their changing relevance to differentdata analysis problems.

To address such tasks as above, a data analysis system has to be equipped with asubstantial amount of background knowledge, and be able to perform symbolic reasoningtasks involving that knowledge and the data. In summary, traditional data analysistechniques facilitate useful data interpretations, and can help to generate important insightsinto the processes behind the data. These interpretations and insights are the ultimateknowledge sought by those who build databases. Yet, such knowledge is not created bythese tools, but instead has to be derived by human data analysts.

In efforts to satisfy the growing need for new data analysis tools that will overcome theabove limitations, researchers have turned to ideas and methods developed in machinelearning. The field of machine learning is a natural source of ideas for this purpose,because the essence of research in this field is to develop computational models foracquiring knowledge from facts and background knowledge. These and related effortshave led to the emergence of a new research area, frequently called data mining and

Page 4: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2

knowledge discovery, (e.g., Michalski, Baskin and Spackman, 1982; Zhuravlev andGurevitch, 1989; Michalski, 1991; Michalski et al, 1992; Van Mechelen et al, 1993;Fayyad et al, 1996; Evangelos and Han, 1996).

The first part of this paper is a compendium of ideas on the applicability of symbolicmachine learning methods to data mining and knowledge discovery. The second partdescribes a methodology for conceptual data exploration, by which we mean the derivationof high-level concepts and descriptions from data. The methodology, stemming mainlyfrom various efforts in machine learning, applies diverse methods and tools for determiningtask-oriented data characterizations and generalizations. These characterizations areexpressed in the form of logic-style descriptions, which can be easily interpreted and usedfor decision-making. The term task-oriented emphasizes the fact that an exploration of thesame data may produce different knowledge; therefore, the methodology tries to connectthe task at hand with the way of exploring the data. Such task-orientation naturally requiresa multistrategy approach, because different tasks may need to employ different dataexploration and knowledge generation operators.

The aim of the methodology is to produce knowledge in a form that is close to datadescriptions that an expert might produce. Such a form may include combinations ofdifferent types of descriptions, e.g., logical, mathematical, statistical, and graphical. Themain constraint is that these descriptions should be easy to understand and interpret by anexpert in the given domain, i.e., they should satisfy the “principle of comprehensibility”(Michalski, 1993). Our first efforts in developing a methodology for multistrategy dataexploration have been implemented in the system INLEN (Michalski et al, 1992). Thesystem combines a range of machine learning methods and tools with more traditional dataanalysis techniques. These tools provide a user with the capability to make different kindsof data explorations and to derive different kinds of knowledge from a database.

The INLEN methodology for intelligent data exploration directly reflects the aims of thecurrent research on data mining and knowledge discovery. In this context, it may be usefulto explain the distinction between the concepts of data mining and knowledge discovery, asproposed in (Fayyad, Piatetsky-Shapiro and Smyth, 1996). According to this distinction,data mining refers to the application of machine learning methods, as well as othermethods, to the “enumeration of patterns over the data,” and knowledge discovery refers tothe process encompassing the entire data analysis lifecycle, from the identification of dataanalysis goals and the acquisition and organization of raw data to the generation ofpotentially useful knowledge, its interpretation and testing. According to these definitions,the INLEN methodology incorporates both data mining and knowledge discoverytechniques.

2 Machine Learning and Multistrategy Data Exploration

This section shows a close relationship between ideas and methods developed in the fieldof machine learning to the goals of data mining and knowledge discovery. Specifically, itdescribes how methods of symbolic machine learning can be used for automating or semi-automating a wide range of tasks concerned with conceptual exploration of data and ageneration of task-oriented knowledge from them. Let us briefly review some of thesemethods.

Page 5: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3

2 . 1 Determining General Rules from Specific Cases

A major class of tools for multistrategy data exploration is based on methods for symbolicinductive learning from examples. Given collections of examples of different decisionclasses (or cases of a relationship), and problem-relevant knowledge (“backgroundknowledge”), an inductive learning method hypothesizes a general description of eachclass. Some methods use a fixed criterion for choosing the description from a largenumber of possibilities, and some allow the user to define a criterion that reflects theproblem at hand. A description can be in the form of a set of decision rules, a decisiontree, a semantic net, etc. A decision rule can also take on many different forms. Here wewill assume the following form:

CLASS ⇐ CONDITION

where CLASS is a statement indicating a class, decision, or a concept name to be assignedto an entity (an object or situation) that satisfies CONDITION; CONDITION is aconjunction of elementary conditions on the values of attributes characterizing the objects;and ⇐ denotes implication.

We will also assume that if CLASS requires a disjunctive description, then several such(conjunctive) rules relate to the same CLASS. To illustrate this point, Figure 1 gives anexample of a disjunctive description of a class of robot-figures in EMERALD (a largesystem for demonstrating machine learning and discovery capabilities—Kaufman andMichalski, 1993).

Rule A: Class 1 ⇐ Jacket Color is Red, Green or Blue &Head Shape is Round or Octagonal

Rule B: Class 1 ⇐ Head Shape is Square &Jacket Color is Yellow

Figure 1. A two-rule description of Class 1.

To paraphrase this description, a robot belongs to Class 1 if the color of its jacket is red,green or blue, and its head is round or octagonal, or, alternatively, its head is square andthe color of its jacket is yellow.

The EMERALD system, mentioned above, combines five programs that display differentkinds of learning capabilities (Kaufman and Michalski, 1993). These capabilities includerule learning from examples (using program AQ15), learning distinctions betweenstructures (INDUCE), conceptual clustering (CLUSTER/2), prediction of object sequences(SPARC), and derivation of equations and rules characterizing data about physicalprocesses (ABACUS). Each of these programs is directly applicable to conceptual dataexploration. For example, the rules in Figure 1 were generated by the AQ15 rule module[MMHL86], [HMM86] from a set of “positive” and “negative” examples of Class 1 ofrobot-figures.

AQ15 learns attributional descriptions of entities, i.e., descriptions involving only theirattributes. More general descriptions, structural or relational., also involve relationshipsamong components of the entities, the attributes of the components, and quantifiers. Suchdescriptions are produced, for example, by the INDUCE module of EMERALD (Larson,1977; Bentrup, Mehler and Riedesel, 1987). Constructing structural descriptions requires

Page 6: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

4

a more complex description language that includes multi-argument predicates, for example,PROLOG, or Annotated Predicate Calculus (Michalski, 1983; Bratko, Muggleton andKaralic, 1997).

For database exploration, attributional descriptions appear to be the most important and theeasiest to implement, because typical databases characterize entities in terms of attributes,not relations. One simple and popular form of attributional description is a decision orclassification tree. In such a tree, nodes correspond to attributes, branches stemming fromthe nodes correspond to attribute values, and leaves correspond to individual classes (e.g.,Quinlan, 1986). A decision tree can be transformed into a set of decision rules (a ruleset)by traversing all paths from the root to individual leaves. Such rules can be oftensimplified by detecting superfluous conditions in them (e.g., Quinlan, 1993). The oppositeprocess of transforming a ruleset into a decision tree is not so direct (Imam, 1995), becausea rule representation is more powerful than a tree representation. The term “morepowerful” means in this context that a decision tree representing a given ruleset may requiresuperfluous conditions (e.g., Michalski, 1990).

The input to an attributional learning program consists of a set of examples of individualclasses and “background knowledge” (BK) relevant to the given learning problem. Theexamples (cases of decisions) are in the form of vectors of attribute-value pairs associatedwith some decision class. Background knowledge is usually limited to information aboutthe legal values of the attributes, their type (the scale of measurement), and a preferencecriterion for choosing among possible candidate hypotheses. Such a criterion may refer to,for example, the computational simplicity of the description, and/or an estimate of itspredictive accuracy. In addition to BK, a learning method may have a representationalbias, i.e., it may constrain the form of descriptions to only a certain type of expressions,e.g., single conjunctions, decision trees, sets of conjunctive rules, or DNF expressions.

In some methods, BK may include more information, e.g., constraints on theinterrelationship between various attributes, rules for generating higher level concepts, newattributes, as well as some initial hypothesis (Michalski, 1983). Learned rules are usuallyconsistent and complete with regard to the input data. This means that they completely andcorrectly classify all the original “training” examples. Sections 5 and 8 present consistentand complete example solutions from the inductive concept learning program AQ15c(Wnek et al, 1995). In some applications, especially those involving learning rules fromnoisy data or learning flexible concepts (Michalski, 1990), it may be advantageous to learndescriptions that are incomplete and/or inconsistent (Bergadano et al, 1992).

Attributional descriptions can be visualized by mapping them into a planar representation ofa discrete multidimensional space (a diagram) spanned over the given attributes (Michalski,1978), (Wnek et al, 1990). For example, Figure 2 shows a diagrammatic visualization ofthe rules from Figure 1. The diagram in Figure 2 was generated by the conceptvisualization program DIAV (Wnek et al, 1990; Wnek, 1995).

Each cell in the diagram represents one specific combination of values of the attributes.For example, the cell marked by an X represents the vector: (HeadShape= S quare,Holding= S word, JacketColor= R ed, IsSmiling= F alse). The four shaded areas markedClass1 (A) represent rule A, and the shaded area marked Class 1 (B) represents rule B. Insuch a diagram, conjunctive rules correspond to certain regular arrangements of cells thatare easy to recognize (Michalski, 1978).

The diagrammatic visualization can be used for displaying the target concept (the conceptto be learned), the training examples (the examples and counter-examples of the concept),

Page 7: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

5

and the actual concept learned by a method. By comparing the target concept with thelearned concept, one can determine the error area, i.e., the area containing all examples thatwould be incorrectly classified by the learned concept. Such a diagrammatic visualizationmethod can illustrate any kind of attributional learning process (Wnek et al, 1990).

X

Class 1

Class 1(A)

Class 1(A)

Class 1(A)

Class 1(A)

Rule A: Class 1 <:: Jacket Color is Red, Green or Blue & Head Shape is Round or Octagonal

Rule B: Class 1 <:: Head Shape is Square & Jacket Color is Yellow

X

(B)

Figure 2. A diagrammatic visualization of rules from Figure 1.

Two types of data exploration operators can be based on methods for learning conceptdescriptions from examples:

• Operators for determining general symbolic descriptions of a designatedgroup or groups of entities in a data set. Such descriptions express thecommon properties of the entities in each group. The operators can useabstract concepts that are not present in the original data via the mechanismof constructive induction (see below). These operators are based onprograms for learning characteristic concept descriptions.

• Operators for determining differences between different groups of entities.Such differences are expressed in the form of rules that define propertiesthat characterize one group but not the other. These operators are based onprograms for learning discriminant concept descriptions.

Page 8: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

6

Section 5 will illustrate these two types of descriptions. For more details and theirdefinitions see (Michalski, 1983). Basic methods for concept learning assume thatexamples do not have errors, that all attributes have a specified value in them, that allexamples are located in the same database, and that concepts to be learned have a precise(“crisp”) description that does not change over time. In many situations one or more ofthese assumptions may not hold. This leads to a variety of more complex machine learningand data mining problems:

• Learning from incorrect data, i.e., learning from examples that contain acertain amount of errors or noise (e.g., Quinlan, 1990; Michalski,Kaufman and Wnek, 1991). These problems are important to learningfrom complex real-world observations, where there is always someamount of noise.

• Learning from incomplete data, i.e., learning from examples in which thevalues of some attributes are unknown (e.g., Dontas, 1988;Lakshminarayan et al, 1996).

• Learning from distributed data, i.e., learning from separate collections ofdata that must be brought together if the patterns within them are to beexposed (e.g., Ribeiro, Kaufman and Kerschberg, 1995).

• Learning drifting or evolving concepts, i.e., learning concepts that are notstable but changing over time, randomly or in a certain general direction.For example, the “area of interest” of a user is often an evolving concept(e.g., Widmer and Kubat, 1996).

• Learning concepts from data arriving over time, i.e., incremental learningin which currently held hypotheses characterizing concepts may need to beupdated to account for the new data (e.g., Maloof and Michalski, 1995).

• Learning from biased data, i.e., learning from a data set that does notreflect the actual distribution of events (e.g., Feelders, 1996).

• Learning flexible concepts, i.e., concepts that inherently lack precisedefinition and whose meaning is context-dependent; some ideas concernedwith this topic include fuzzy sets (e.g., Zadeh, 1965; Dubois, Prade andYager, 1993), two-tiered concept representations (e.g., Michalski, 1990;Bergadano et al, 1992), and rough sets (e.g., Pawlak, 1991; Slowinski,1992; Ziarko, 1994).

• Learning concepts at different levels of generality, i.e., learningdescriptions that involve concepts from different levels of generalizationhierarchies representing background knowledge (e.g., Kaufman andMichalski, 1996).

• Integrating qualitative and quantitative discovery, i.e., determining sets ofequations that fit a given set of data points, and qualitative conditions forthe application of these equations (e.g., Falkenhainer and Michalski,1990).

• Qualitative prediction, i.e., discovering patterns in sequences or processesand using these patterns to qualitatively predict the possible continuation ofthe given sequences or processes (e.g., Davis, 1981; Michalski, Ko andChen, 1985; 1986; Dieterrich and Michalski, 1986).

Each of these problems is relevant to the derivation of useful knowledge from a collectionof data (static or dynamic). Therefore, methods for solving these problems developed in

Page 9: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

7

the area of machine learning are directly relevant to data mining and knowledge discovery,in particular, to conceptual data exploration.

2 . 2 Conceptual Clustering

Another class of machine learning methods relevant to data mining and knowledgediscovery concerns the problem of building a conceptual classification of a given set ofentities. The problem is similar to that considered in traditional cluster analysis, but isdefined in a different way. Given a set of attributional descriptions of some entities, adescription language for characterizing classes of such entities, and a classification qualitycriterion, the problem is to partition entities into classes in a way that maximizes theclassification quality criterion, and simultaneously to determine general (extensional)descriptions of these classes in the given description language. Thus, a conceptualclustering method seeks not only a classification structure of entities (a dendrogram), butalso a symbolic description of the proposed classes (clusters). An important,distinguishing aspect of conceptual clustering is that, unlike in cluster analysis, theproperties of class descriptions are taken into consideration in the process of determiningthe classes (clusters).

To clarify the difference between conceptual clustering and conventional clustering, noticethat a conventional clustering method typically determines clusters on the basis of asimilarity measure that is a function solely of the properties (attribute values) of the entitiesbeing compared, and not of any other factors:

Similarity(A, B) = f(properties(A), properties(B))where A and B are entities being compared.

In contrast, a conceptual clustering program clusters entities on the basis of a conceptualcohesiveness, which is a function of not only properties of the entities, but also of twoother factors: the description language L, which the system uses for describing the classesof entities, and of the environment, E, which is the set of neighboring examples:

Conceptual cohesiveness(A, B) = f(properties(A), properties(B), L, E)

A

B

Figure 3. An illustration of the difference between closeness and conceptual cohesiveness.

Thus, two objects may be similar, i.e., close according to some distance (or similarity)measure, while having a low conceptual cohesiveness, or vice versa. An example of thefirst situation is shown in Figure 3. The points (black dots) A and B are “close” to each

Page 10: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

8

other; they therefore would be placed into the same cluster by any technique based solelyupon the distances between the points. However, these points have small conceptualcohesiveness due to the fact that they belong to configurations representing differentconcepts. A conceptual clustering method, if equipped with an appropriate descriptionlanguage, would cluster the points in Figure 3 into two “ellipses,” as people normallywould.

A classification quality criterion used in conceptual clustering may involve a variety offactors, such as the fit of a cluster description to the data (called sparseness), the simplicityof the description, and other properties of the entities or the concepts that describe them(Michalski, Stepp and Diday, 1981). An example of conceptual clustering is presented inSection 5.

Some new ideas on employing conceptual clustering for structuring text databases andcreating concept lattices for discovering dependencies in data are in (Carpineto andRomano, 1995a; 1995b). The concepts created through the clustering are linked in latticestructures that can be traversed to represent generalization and specialization relationships.

2 . 3 Constructive Induction

Most methods for learning rules or decision trees from examples assume that the attributesused for describing examples are sufficiently relevant to the learning problem at hand. Thisassumption does not always hold in practice. Attributes used in the examples may not bedirectly relevant, and some attributes may be irrelevant or nonessential. An importantadvantage of symbolic methods over statistical methods is that they can relatively easilydetermine irrelevant or nonessential attributes. An attribute is nonessential if there is acomplete and consistent description of the classes or concepts to be learned that does notuse this attribute. Thus, a nonessential attribute may be either irrelevant or relevant, butwill by definition be dispensable. Inductive learning programs such as the rule-learningprogram AQ, or the decision tree-learning ID3, can cope relatively easily with a largenumber of nonessential attributes in their input data.

If there are very many nonessential attributes in the initial description of the examples, thecomplexity of a learning process may significantly increase. Such a situation calls for amethod that can efficiently determine the most relevant attributes for the given problemfrom among all those given initially. Only the most relevant attributes will be used in thedescription learning process. Determining the most relevant attributes is therefore a usefuldata exploration operator. Such an operator can also be useful for the data analyst on itsown merit, as it may be important to know which attributes are most discriminatory for agiven set of classes. By removing less relevant attributes, the representation space isreduced, and the problem becomes simpler. Thus, such a process can be viewed as a formof improving the representation space. Some methods for finding the most relevantattributes are described in (Zagoruiko, 1972; Baim, 1982).

In many applications, the attributes originally given may be only weakly or indirectlyrelevant to the problem at hand. In such situations, there is a need for generating new,more relevant attributes that may be functions of the original attributes. These functionsmay be simple, e.g., a product or sum of a set of the original attributes, or very complex,e.g., a Boolean attribute based on the presence or absence of a straight line or circle in animage (Bongard, 1970). Finally, in some situations, it will be desirable to abstract someattributes, that is, to group some attribute values into units, and thus reduce the attribute’s

Page 11: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

9

range of possible values. A quantization of continuous attributes is an example of such anoperation.

All the above operations—removing less relevant attributes, adding more relevantattributes, and abstracting attributes—are different forms of improving the originalrepresentation space for learning. A learning process that consists of two (intertwined)phases, one concerned with the construction of the “best” representation space, and thesecond concerned with generating the “best” hypothesis in the found space is calledconstructive induction (Michalski, 1978; 1983; Wnek and Michalski, 1994). An exampleof a constructive induction program is AQ17 (Bloedorn, Wnek and Michalski, 1993),which performs all three types of improvements of the original representation space. Inthis program, the process of generating new attributes is done by combining initialattributes by mathematical and/or logical operators and selecting the “best” combinations,and/or by obtaining advice from an expert (Bloedorn, Wnek and Michalski, 1993;Bloedorn and Michalski, 1996).

2 . 4 Selection of the Most Representative Examples

When a database is very large, determining general patterns or rules characterizing differentconcepts may be very time-consuming. To make the process more efficient, it may beuseful to extract from the database the most representative or important cases (examples) ofgiven classes or concepts. Most such cases are those that are either most typical or mostextreme (assuming that there is not too much noise in the data). One method fordetermining the latter ones, the so-called “method of outstanding representatives,” isdescribed in (Michalski and Larson, 1978).

2 . 5 Integration of Qualitative and Quantitative Discovery

In a database that contains numerical attributes, a useful discovery might be an equationbinding these attributes. For instance, from a table of planetary data including planets’masses, densities, distances from the sun, periods of rotation, and lengths of local years,one could automatically derive Kepler’s Law that the cube of the planet’s distance from thesun is proportional to the square of the length of its year. This is an example of quantitativediscovery. The application of machine learning to quantitative discovery was pioneered bythe BACON system (Langley, Bradshaw and Simon, 1983), and then explored by manysystems since, such as COPER (Kokar, 1986), FAHRENHEIT (Zytkow, 1987), andABACUS (Falkenhainer and Michalski, 1990). Similar problems have been exploredindependently by Zagoruiko (1972) under the name of empirical prediction.

Some equations may not apply directly to data, because of an inappropriate value of aconstant, or different equations may apply under different qualitative conditions. Forexample, in applying Stoke’s Law to determine the velocity of a falling ball, if the ball isfalling through a vacuum, its velocity depends on the length of time it has been falling andon the gravitational force being exerted upon it. A ball falling through some sort of fluidwill reach a terminal velocity dependent on the radius and mass of the ball and the viscosityof the fluid.

A program ABACUS (Greene, 1988; Falkenhainer and Michalski, 1990; Michalski, 1991)is able to determine quantitative laws under different qualitative conditions. It partitions thedata into example sets, each of which adheres to a different equation determined by aquantitative discovery module. The qualitative discovery module can then determine

Page 12: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 0

conditions/rules that characterize each of these example sets (in the case of Stoke’s Law,the rules would be based on the medium of descent).

2 . 6 Qualitative Prediction

Most programs that determine rules from examples determine them from instances ofvarious classes of objects. An instance of a concept exemplifies that concept regardless ofits relationship to other examples. Contrast that with a sequence prediction problem, inwhich a positive example of a concept is directly dependent on the position of the examplein the sequence. For example, Figure 4 shows a sequence of seven figures. One may askwhat object plausibly follows in the eighth position? To answer such a question, one needsto search for a pattern in the sequence, and then use the pattern to predict a plausiblesequence continuation. In qualitative prediction, the problem is not to predict a specificvalue of a variable (as in time series analysis), but to qualitatively describe a plausiblefuture object, that is, to describe plausible properties of a future object.

1 2 3 4 5 6 7

?

Figure 4. An example of a sequence prediction problem.

In the example in Figure 4, one may observe that the sequence consists of T-shaped figureswith black tips and I-shaped figures with white tips. The figures may be white or shaded,and may be rotated in different orientations at 45-degree intervals. But is there a consistentpattern?

To determine such a pattern, one can employ different descriptive models, and instantiatethe models to fit the particular sequence. The instantiated model that best fits the data isthen used for prediction. Such a method is described in (Dieterrich and Michalski, 1986).The method employs three descriptive models—periodic, decomposition and DNF.

The periodic model is used to detect repeating patterns in a sequence. For example, Figure4 depicts a recurring pattern that alternates T-shaped and I-shaped objects. In general, therecan also be periodic sequences within the periodic sequences. In the figure, the T-shapedobjects form a subsequence in which individual objects rotate leftward by 45 degrees.

The second model, the decomposition model, is used to characterize a sequence by decisionrules in the following general form: “If one or more of the previous elements of thesequence have a given set of characteristics, then the next element will have the followingcharacteristics.” One such rule that applies to the sequence in Figure 4 would state that ifan element in the sequence has a vertical component, then the next element in the sequencewill have a shaded component; otherwise it will have no shaded components.

The third model, the DNF (disjunctive normal form) or “catch-all” model, tries to capturegeneral properties characterizing the whole sequence. For example, for the sequence inFigure 4, it could instantiate to a statement such as “all elements in the sequence are T-shaped or I-shaped, they have white or shaded interiors, white or black tips, etc.

Page 13: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 1

The program SPARC/G (Michalski, Ko and Chen, 1986) employs these three descriptivemodels to detect patterns in a sequence of arbitrary objects, and then uses the patterns topredict a plausible continuation for the sequence. For the sequence in Figure 4, SPARC/Gfound the following strong pattern based on the periodic model:

Period< [Shape=T-shape] & [orientation(i+1)=orientation(i) - 45],[Shape = I-shape] & [orientation(i+1)=orientation(i) + 45] &[shaded(i+1)=unshaded(i)]>

The pattern can be paraphrased: There are two phases in a repeating period (theirdescriptions are separated by a comma). The first phase involves a T-shaped figure, andthe second phase an I-shaped figure. The T-shaped figure rotates to the left, and the I-shaped figure rotates to the right by 45 degrees in relation to its predecessor. I-shapedfigures are alternatingly shaded and unshaded. Based on this pattern, a plausible nextfigure in the sequence would be an unshaded I-shaped figure rotated clockwise 45 degreesin relation to the previous I-shaped figure.

The qualitative prediction capabilities described above can be useful for conceptualexploration of temporal databases in many application domains, such as agriculture,medicine, robotics, economic forecasting, etc.

2 . 7 Summarizing the Machine Learning-Oriented Approach

To help the reader to develop a rough sense of what is different and new in the above, letus summarize operations typically performed by traditional multivariate data analysismethods. These include computing mean-corrected or standardized variables, variances,standard deviations, covariances and correlations among attributes; principal componentanalysis (determining orthogonal linear combinations of attributes that maximally accountfor the given variance); factor analysis (determining highly correlated groups of attributes);cluster analysis (determining groups of data points that are close according to some distancemeasure); regression analysis (fitting an equation of an assumed form to given data points);multivariate analysis of variance; and discriminant analysis. All these methods can beviewed as primarily oriented toward a numerical characterization of a data set.

In contrast, the machine learning methods described above are primarily oriented towarddeveloping symbolic logic-style descriptions of data, which may characterize one or moresets of data qualitatively, differentiate between different classes (defined by different valuesof designated output variables), create a “conceptual” classification of data, select the mostrepresentative cases, qualitatively predict sequences, etc. These techniques are particularlywell suited for developing descriptions that involve nominal (categorical) and rank variablesin data.

Another important distinction between the two approaches to data analysis is that statisticalmethods are typically used for globally characterizing a class of objects (a table of data), butnot for determining a description for predicting class membership of future objects. Forexample, a statistical operator may determine that the average lifespan of a certain type ofautomobile is 7.3 years. Knowledge of the average lifespan of automobiles in a given classdoes not allow one to recognize the type of a particular automobile for which one obtainedinformation about how long this automobile remained driveable. In contrast, a symbolicmachine learning approach might create a description such as “if the front height of avehicle is between 5 and 6 feet, and the driver’s seat is 2 to 3 feet above the ground, then

Page 14: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 2

the vehicle is likely to be a minivan.” Such descriptions are particularly suitable forassigning entities to classes on the basis of their properties.

The INLEN methodology integrates a wide range of strategies and operators for dataexploration based on machine learning research, as well as statistical operators. The reasonfor such a multistrategy approach is that a data analyst may be interested in many differenttypes of information about the data. Different types of questions require differentexploratory strategies and different operators.

3 Classification of Data Exploration Tasks

The problems described above can be simply illustrated by means of a general data table(GDT). Such a table is a generalization of a standard data table used in data analysis(Figure 5). It consists of a collection of relational tables (data tables) arranged in layersordered by the time instance associated with each table. A GDT is used to represent asequence of entities as they change over time. Examples of a GDT are a sequence ofmedical records of a patient (when each record is represented as a table of test results), asequence of descriptions of a crop as it develops in the field, a sequence of data tablescharacterizing the state of a company during selected time instances, etc.

Selectingmost representativeexamples

Selecting most relevantattributes

Tuples

. . .

. . .

. . .

. . .

...

...

1

1

A0 A1 Ai An+1 An+2

Original Attributes Generated Attributes

ClassA0 = 1extension

ClassA0 = kextension

Time … …

kk

Examples of Class A0 = 11

Examples of Class A0 = k

An

Figure 5. A GDT illustrating the role of different symbolic operators.

Columns in the tables correspond to attributes used to characterize entities associated withthe rows. These may be initial attributes, given a priori, or additional ones generatedthrough a process of constructive induction (e.g., Wnek and Michalski, 1994). Each

Page 15: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 3

attribute is assigned a domain and a type. The domain specifies the set of all legal valuesthat the attribute can be assigned in the table. The type defines the ordering (if any) of thevalues in the domain. For example, the AQ15 learning program (Michalski et al, 1986)allows four types of attributes: nominal (no order), linear (total order), cyclic (cyclic totalorder), and structured (hierarchical order; see Kaufman and Michalski, 1996). Theattribute type determines the kinds of operations that are allowed on this attribute’s valuesduring a learning process.

Entries in each row are values of the attributes for the entity associated with the row.Typically, each row corresponds to a single entity. However, in large databases whoserecords represent common, repeatable transactions, a column can be added to represent thenumber of occurrences of that particular transaction. With such information, discoverytools can incorporate a bias based on the frequency of instances.

Entries in the various columns of the table can be specific values of the correspondingattributes, the symbol “?,” meaning that a value of this attribute is unknown for the givenentity, or the symbol N/A, if an attribute does not apply to a specific entity. For example,“number of legs” usually applies to an animal, but would not apply to a plant.

An important problem of conceptual data exploration is to determine which attribute orattributes in a table functionally depend on other attributes. A related problem is todetermine a general form of this relationship that would enable one to predict values ofsome attributes for future entities. For instance, when it is known that a nominal-scaleattribute depends on other (independent) attributes, the problem is to hypothesize a generaldescription of this relationship so that one can predict values of the nominal-scale attributefor future combinations of values of the independent attributes. This problem is equivalentto the problem of concept learning from examples, so methods developed in machinelearning directly apply. In such a case, the column in the data table that corresponds to thedependent attribute represents the output attribute. The values of that variable are classeswhose descriptions are to be learned. In Figure 5, for illustration, it was assumed that thefirst column (attribute A0) represents values of the output variable. When there are no apriori classes to which entities belong, there is no such designated column. In this case,methods of conceptual clustering can be applied to determine a classification of entities.

Below we use the GDT (Figure 5) to relate machine learning techniques described in theprevious section to data exploration problems.

Learning rules from examples:

Suppose that one discrete attribute in the GDT has been designated as the output attribute,and all or some of the remaining attributes as input (independent) attributes. A set of rowsin the table for which the output attribute takes the same value can be viewed as a set oftraining examples of the decision class (concept) symbolized by this value. Any of theconventional concept learning techniques can be directly applied for determining a rulerelating the output attribute to the input attributes. For a general analysis of the data set,every discrete attribute (and continuous attributes as well after quantization) can beconsidered as an output attribute, and a machine learning method can be applied todetermine a relationship between that attribute and other attributes. The determination ofsuch relationships (rules) can be guided by different rule quality criteria, for example,simplicity, cost, predictive accuracy, etc. In the INLEN system, the AQ learning methodwas applied due to the simplicity and the high comprehensibility of decision rules itgenerates (Wnek et al, 1995; Bloedorn and Michalski, 1996).

Page 16: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 4

Determining time-dependent patterns:

This problem concerns the detection of temporal patterns in sequences of data arrangedalong the time dimension in a GDT (Figure 5). Among the novel ideas that could beapplied for analyzing such time-dependent data is a multi-model method for qualitativeprediction (Dieterrich and Michalski, 1986; Michalski, Ko and Chen, 1985; 1986).Another novel idea is a temporal constructive induction technique that can generate newattributes that are designed to capture time-dependent patterns (Davis, 1981; Bloedorn andMichalski, 1996).

Example selection:

The problem is to select rows from the table that correspond to the most representativeexamples of different classes. When a datatable is very large, is it important to concentratethe analysis on a representative sample. The “method of outstanding representatives”selects examples (tuples) that are most different from the other examples (Michalski andLarson, 1978).

Attribute selection:

When there are many columns (attributes) in the GDT, it is often desirable to reduce thedata table by removing columns that correspond to the least relevant attributes for adesignated learning task. This can be done by applying one of many methods for attributeselection, such as Gain Ratio(Quinlan, 1993) or Promise level (Baim, 1982).

Generating new attributes:

The problem is to generate additional columns that correspond to new attributes generatedby a constructive induction procedure. These new attributes are created by using theproblem’s background knowledge and/or special heuristic procedures as described inpapers on constructive induction (e.g., Bloedorn, Wnek and Michalski, 1993).

Clustering:

The problem is to automatically partition the rows of the table into groups that correspondto “conceptual clusters,” that is, sets of entities with high conceptual cohesiveness(Michalski, Stepp and Diday, 1981). Such a clustering operator will generate an additionalcolumn in the table that corresponds to a new attribute ”cluster name.” The values of thisattribute for each tuple in the table indicate the assigned class of the entity. Rules thatdescribe clusters are stored separately in the Knowledge Base and linked to the entities viaknowledge segments (see Section 4). An example of a clustering is presented in Section 5.

Determining attribute dependencies:

The problem is to determine relationships, such as correlations, causal dependencies,logical or functional dependencies among the attributes (columns) in the given GDT, usingstatistical and logical methods.

Incremental rule update:

The problem is to update working knowledge (in particular, rulesets characterizingrelationships among attributes in the GDT) to accommodate new instances or time slices inthe table. To do so, an incremental learning program must be applied to synthesize theprior knowledge with the new information. The incremental learning process may be full-memory, partial-memory, or no-memory, depending on how much of the original trainingdata is maintained in the incremental learning process (Hong, Mozetic and Michalski, 1986;Reinke and Michalski, 1988; Maloof and Michalski, 1995).

Page 17: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 5

Searching for approximate patterns in (imperfect) data:

For some GDTs, it may not be possible (or useful) to find complete and consistentdescriptions. In such cases, it is important to determine patterns that hold for a largenumber of cases, but not necessarily for all. An important case of this problem is whensome entries in the table are missing or incorrect. The problem is then to determine the best(i.e., the most plausible) hypothesis that accounts for most of the available data.

Filling in missing data:

Given a data table in which some entries are missing, determine plausible values of themissing entries on the basis of analysis of the currently known data. An interestingapproach to this problem is to apply a multi-line reasoning, based on the core theory ofhuman plausible reasoning (Collins and Michalski, 1981; 1989; Dontas, 1988).

Determining decision structures from declarative knowledge (decision rules):

Suppose that a set of general decision rules (a declarative form of knowledge) has beenhypothesized for a given data set (GDT). If this ruleset is to be used for predicting newcases (by a computer program, or by an expert), it may be desirable to convert it into theform of a decision tree (or a more general form, a decision structure) that is tailored to agiven decision-making situation (e.g., by taking into consideration the cost of measuringattributes). A methodology for doing this and arguments for and against using such anapproach (as opposed to the traditional method of learning of decision trees directly fromexamples) are discussed in (Imam and Michalski, 1993; Imam ,1995; Michalski and Imam,1997).

Methods for performing the above operations on data tables have been implemented invarious machine learning programs (e.g., Michalski, Carbonell and Mitchell, 1983; 1986;Forsyth and Rada, 1986; Kodratoff, 1988; Kodratoff and Michalski, 1990). Below wedescribe the INLEN system that aims at ultimately incorporating all of these programs asoperators in one integrated system for the generation of knowledge from data.

4 Integration of Many Operators in INLEN

To make the data exploration operations described above easily available to a data analyst,and applicable in sequences in which the output from one operation is an input to anotherone, programs performing these operations need to be integrated into one system. Thisidea underlies the INLEN system (Kaufman, Michalski and Kerschberg, 1991; Michalskiet al, 1992). The name INLEN is derived from inference and learning. The systemintegrates machine learning programs, statistical data analysis tools, a database, aknowledge base, inference procedures, and various supporting programs under a unifiedarchitecture and graphical interface. The knowledge base is used for storing, updating andapplying rules and other forms of knowledge that may be employed for assisting dataexploration, and for reporting results from it.

The general architecture of INLEN is presented in Figure 6. The system consists of adatabase (DB) connected to a knowledge base (KB), and a set of operators. The operatorsare divided into three classes:

• DMOs: Data Management Operators, which operate on the database.These are conventional data management operators that are used forcreating, modifying and displaying relational tables.

Page 18: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 6

• KMOs: Knowledge Management Operators, which operate on theknowledge base. These operators play a similar role to the DMOs, butapply to the rules and other structures in the knowledge base.

• KGOs: Knowledge Generation Operators, which operate on both the dataand knowledge bases. These operators perform symbolic and numericaldata exploration tasks. They are based on various machine learning andinference programs, on conventional data exploration techniques, and onvisualization operators for displaying graphically the results of exploration.The diagrammatic visualization method DIAV (Wnek ,1995) is used fordisplaying the effects of symbolic learning operations on data.

DB KB

SELECT

PROJECT

JOIN

COMBINE

CREATE

INSERT

CHANGE

DELETE

INTERSECT

SELECT

PROJECT

JOIN

COMBINE

INTERSECT

CREATE

INSERT

CHANGE

DELETE

DATA MANAGEMENT OPERATORS

KNOWLEDGEMANAGEMENT OPERATORS

VISUALIZE

GENTREE GENHIER

GENEQGENRULE TRANSFORM

ANALYZE GENATR

TEST GENEVE

KNOWLEDGE GENERATION OPERATORS

KNOWLEDGE GENERATION OPERATORS

Figure 6. An architecture of the INLEN system for multistrategy data exploration.

The KGOs are the heart of the INLEN system. To facilitate their use, the concept of aknowledge segment was introduced (Kaufman, Michalski and Kerschberg, 1991). Aknowledge segment is a structure that links one or more relational tables from the databasewith one or more structures from the knowledge base. KGOs can be viewed as modulesthat perform some form of inference or transformation on knowledge segments and, as aresult, create new knowledge segments. Knowledge segments are both inputs to andoutputs from the KGOs. Thus, they facilitate the passage of data and knowledge from oneknowledge generation operator to another.

The execution of a KGO usually requires some background knowledge, and is guided bycontrol parameters (if some parameters are not specified, default values are used). Thebackground knowledge may contain some general knowledge as well as knowledgespecifically relevant to a given application domain, such as a specification of the value setsand types of attributes, the constraints and relationships among attributes, initial rules

Page 19: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 7

hypothesized by an expert, etc. The KGOs can be classified into groups, based on the typeof operation they perform. Each group includes a number of specific operators that areinstantiated by a combination of parameters. The basic operator groups are as follows:

• GENRULE operators generate different kinds of decision rules from given facts. Aspecific operator may generate rules characterizing a set of facts, discriminatingbetween groups of facts, characterizing a sequence of events, and determiningdifferences between sequences, based on programs such as AQ15c (Wnek et al, 1995)and SPARC/G (Michalski, Ko and Chen, 1986). A KGO for learning rules canusually work in either incremental or batch mode. In the incremental mode, it tries toimprove or refine the existing knowledge, while in the batch mode, it tries to createentirely new knowledge based on the facts in the database, and knowledge in theknowledge base.

• GENTREE operators build a decision structure from a given set of decision rules (e.g.,Imam and Michalski, 1993), or from examples (e.g., Quinlan, 1993). A decisionstructure is a generalization of the concept of a decision tree in which nodes can beassigned an attribute or a function of attributes. Individual branches may be assigned aset of attribute values. Leaves may be assigned a set of decisions (Imam andMichalski, 1993; Imam, 1995).

• GENEQ operators generate equations characterizing numerical data sets andqualitatively describing the conditions under which these equations apply (e.g.,Falkenhainer and Michalski, 1990).

• GENHIER operators build conceptual clusters or hierarchies. They are based on theprogram CLUSTER methodology (Michalski, Stepp and Diday, 1981). The operatorin INLEN is based on the reimplementation in C of the program CLUSTER/2 (Stepp,1984).

• TRANSFORM operators perform various transformations on the knowledge segments,e.g., generalization or specialization, abstraction or concretion, optimization of givenrules, etc. according to user-provided criteria. For instance, one such operator climbsan attribute’s generalization hierarchy to build more general decision rules (Kaufmanand Michalski, 1996).

• GENATR operators generate new attribute sets by creating new attributes (Bloedornand Michalski, 1996), selecting the most representative attributes from the original set(Baim, 1982), or by abstracting attributes (Kerber, 1992).

• GENEVE operators generate events, facts or examples that satisfy given rules, selectthe most representative events from a given set (Michalski and Larson, 1978),determine examples that are similar to a given example (Collins and Michalski, 1989),or predict the value of a given variable using an expert system shell or a decisionstructure.

• ANALYZE operators analyze various relationships that exist in the data, e.g.,determining the degree of similarity between two examples, checking if there is animplicative relationship between two variables, etc. Statistical and symbolic operatorsalike may perform these tasks.

Page 20: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 8

• TEST operators test the performance of a given set of rules on an assumed set of facts.The output from these operators is a confusion matrix—a table whose (i,j)th elementshows how many examples from the class i were classified by the rules to be in class j.These operators can also be used to apply the rules to any given situation to determine adecision. The TEST operator implemented in INLEN is based on the ATEST program(Reinke, 1984).

• VISUALIZE operators are used to present data and/or knowledge to the user in aconvenient, easy-to-understand format (Wnek, 1995).

Summarizing, INLEN integrates a large set of operators for performing various types ofoperations on the data base, on the knowledge base, or the data and knowledge basescombined.

5 Illustration of Clustering and Learning Operators

Among the most important knowledge generation operators implemented in INLEN are theoperator for creating a classification of data (clustering), and the operator for learninggeneral rules relating a designated concept (attribute) to other designated attributes. Thefirst operator is realized by the CLUSTER/2 program for conceptual clustering (Stepp,1984). The second operator is realized by the AQ15c rule learning program (Wnek et al,1995). This section illustrates these operators through an application to a datatablecharacterizing hard drives (Figure 7). The datatable is based on information published inthe October, 1994 issue of MacUser.

In the table presented in Figure 7, each row (except for the first one) describes a hard drivein terms of the attributes specified in the first row. Suppose that the task of dataexploration is to develop a classification of the hard drives into some meaningfulcategories. For this task, the operator CLUSTER is applied. Let us assume that theoperator will seek a clustering that maximizes the quality of classification, as defined bytwo criteria: the simplicity of the descriptions of generated categories, and the cohesivenessof the descriptions (measured by the ratio of the number of instances in the datatablecovered by a given description to the number of possible instances covered by thedescription). The input to the conceptual clustering operator is the table in Figure 7(without the rightmost column, which, for the sake of saving space, represents already theresult of clustering).

The result of applying the clustering operator is a knowledge segment containing twocomponents—a new, extended data table, and a set of rules. The new table, in comparisonto the input table, has an additional column—the rightmost column in Figure 7, labeled“Group,” which represents the category assignments of the drives by the clusteringoperator. The second component is the set of rules describing the categories that weregenerated. Here are the rules describing the categories created by the operator:[Class 1] ⇐ [Toll_free_Support is yes] & [FCC_Class-B is yes] & [Encryption is

no] & [SCSI_50-Pin is yes or no] & [Guarantee is yes or by dealer][Class 2] ⇐ [Toll_free_Support is no] & [SCSI_50-Pin is yes] & [5yr_Warranty

is yes] & [Guarantee is yes or no] & [Loaners is yes or no][Class 3] ⇐ [Toll_free_Support is no] & [FCC_Class-B is yes] & [AC outlet is

yes] & [Passwd_Protect is yes] & [5yr_Warranty is no] & [Guaranteeis not by dealer] & [Loaners is yes or if available]

Page 21: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

1 9

Hard Drive

Apple 1050

Micropolis

SLMO 1000

Focus 1G

GHD 1200S

Joule 1080

Liberty 1GB

Spitfire 1GB

PowerUser1070P1000

Seagate 1075

Minipak 1000

PowerCity1GBSpin 1021

APS MS 1.7

Seagate 2GB

SLMO 2000

Focus 2G

FWB 1760MF

Liberty 2GB

Loviel L2000

Seagate 2.1

PowerUser1801MacP Sg 28

ACOutlet

SCSI50-Pin

FCCClass B

PasswdProtectEncrypt

5yrWarranty

Toll-freeSupportGuaranteeLoanersCapacity

no

no

no

yes

no

yes

no

yes

no

no

yes

no

yes

no

no

no

no

yes

no

no

yes

yes

no

no

yes

yes

yes

yes

yes

no

25 pin

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

68 pinSCSI2

no

yes

yes

yes

yes

yes

yes

ClassA

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

no

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

no

yes

yes

yes

no

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

no

yes

no

yes

no

no

no

no

no

no

no

no

no

no

no

no

no

yes

no

no

yes

no

no

no

no

no

no

yes

yes

no

yes

yes

no

yes

no

no

no

no

onmech.yes

yes

yes

yes

no

no

no

yes

yes

no

yes

yes

no

no

yes

no

yes

no

no

yes

yes

yes

no

yes

yes

yes

no

no

yes

no

no

no

no

yes

no

bydealer

no

no

yes

no

yes

yes

yes

yes

yes

yes

yes

yes

yes

yes

no

no

yes

no

yes

yes

yes

yes

yes

low

low

low

low

low

low

low

low

low

low

low

low

low

low

high

high

high

high

high

high

high

high

high

high

bydealer

no

yes

yes

no

no

yes

no

no

no

no

yes

no

no

no

no

yes

yes

ifavail.yes

yes

no

no

no

Group

1

2

2

1

2

1

3

2

1

1

1

3

1

1

1

2

2

1

3

3

2

2

1

2

Figure 7. A datatable characterizing hard drives.

Thus, the operator created three categories of hard drives and described each category in theform of rules. Each rule shows all characteristics common to a given category, that is, itrepresents a characteristic description of a category (Michalski, 1983). (Note that some ofthe conditions in these rules appear to be redundant. For example, the last condition of theClass 2 rule says that Loaners is yes or no. This can be explained by the presence of athird value, “by dealer,” that neither guarantees, nor rules out a loaner.) Thesecharacterizations do not point out the most significant distinctions between a given categoryand other categories.

To create a description that points out the most significant distinctions, one needs to applythe operator that creates discriminant descriptions (Michalski, 1983). The operator(GENRULE) is applied to the extended data table in Figure 7, using the “Group” columnas its output attribute. The result is a set of new decision rules:[Class 1] ⇐ [Toll_free_Support is yes]

[Class 2] ⇐ [Toll_free_Support is no] & [5yr_Warranty is yes]

[Class 3] ⇐ [Toll_free_Support is no] & [5yr_Warranty is no]

Page 22: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 0

The obtained rules are much simpler and easier to interpret than the rules generated by theCLUSTER operator that invented the three classes. The reason is that a discriminantdescription lists only those characteristics that are necessary to discriminate a givencategory from the other categories. Discriminant descriptions are designed to provide theminimum information needed for distinguishing between entities of different categories.Both characteristic and discriminant descriptions are complete and consistent with all theexamples in Figure 7, i.e., they classify all examples in the same way.

6 Data and Rule Visualization

It is desirable for data analysts to be able to visualize the results of different operators inorder to relate visually the input data to the rules that have been learned from them, to seewhich datapoints would corroborate or contradict these rules, to identify possible errors,etc. To this end, INLEN supports the visualization of data and knowledge through thediagrammatic visualization method implemented in the DIAV program (Michalski, 1978;Wnek, 1995).

Figure 8. A visualization of the characteristic description created by the conceptualclustering operator.

Let us illustrate the method with the hard disk classification problem presented in theprevious section. The representation space, projected onto six attributes, is pictured inFigure 8. To simplify the visualization, the attributes used to span the diagram,Toll_free_Support (tf), Loaners (lo), SCSI_50-Pin (sc), FCC_Class-B (fc), Guarantee(gu), and 5yr_Warranty (wa), are only those that appeared most frequently in thecharacteristic descriptions created by the conceptual clustering operator. Each cell in thediagram corresponds to one combination of attribute values, specified by the annotations of

Page 23: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 1

the columns and rows. Thus the upper-leftmost cell corresponds to a datapoint in which allsix of these attributes have the value yes (y).

The 24 examples from Figure 7 have been projected onto this space, and are represented byplacing their class number in the corresponding cells. The shaded areas represent thecharacteristic descriptions of the classes generated by the clustering operator; the lightestcolor indicates Class 1, the intermediate shade represents Class 2, and the darkest oneindicates Class 3. As can be seen in the diagram, the descriptions generated by theclustering operator are generalizations of the input instances, as they also cover instancesthat have not yet been observed (shaded areas without a number).

For comparison, Figure 9 is a visualization of the discriminant descriptions generated bythe rule learning operator from the input examples classified according to the previouslygenerated clustering. The organization of the diagram in Figure 9 is the same as in Figure 8with regard to the labeling of examples, classes, rows and columns. Because discriminantdescriptions focus only on features that distinguish among the classes, they cover broadersections of the representation space. Thus, they are much more general than characteristicdescriptions.

Figure 9. A visualization of the discriminant rules created by the inductive generalizationoperator.

The obtained discriminant descriptions divide the representation space into four sections,three corresponding to the rules for the three classes, and the fourth to the indeterminateportion of the event space, containing none of the known instances of the three categories.This latter section is defined by the combination of characteristics: Toll_free_Support = noand 5yr_Warranty = on_mechanism.

Page 24: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 2

Note also that due to the greater generality of the discriminant descriptions, theindeterminate area is much smaller than in the case of characteristic descriptions (the blankarea in Figure 8).

As can be seen from the diagram, the generated discriminant descriptions are consistent andcomplete with regard to all of the presented examples, that is, they preserve theclassification of cases created by the clustering operator. Summarizing, the visualizationmethod presented above makes it very easy to see how generated descriptions relate to thecases from which they were generated.

7 Learning Rules with Structured Attributes

In addition to conventional symbolic and numerical attributes, INLEN supports a new kindof attribute, called structured. Such attributes have value sets ordered into hierarchies(Michalski, 1980). In order to take advantage of the properties of structured attributes inexecuting inductive learning, new inductive generalization rules have been defined.

An inductive generalization rule (or transmutation) takes an input statement and relevantbackground knowledge, and hypothesizes a more general statement (Michalski, 1980;1983; 1994). For example, removing a condition from the premise of a decision rule is ageneralization transmutation (this is called a dropping condition generalization rule), since ifthe premise has fewer conditions, a larger set of instances can satisfy it.

A powerful inductive generalization operator used in the AQ learning programs is theextension-against operator. If rule R1: C ⇐ [xi = A] & CTX1 characterizes a subset of

positive concept examples, E+, of the concept C, and rule R2: C ⇐ [x i = B] & CTX2characterizes negative examples, E- (where A and B represent disjoint subsets of the valuesof xi, and the CTXs stand for any additional conditions), then the extension of R1 againstR2 along dimension xi

C ⇐ R1 | R2 /x i

produces a new rule R3: [x i ≠ B ∪ ε], which is a consistent generalization of R1, that is,a generalization that does not intersect logically with R2 (Michalski and McCormick, 1971;Michalski, 1983). The value of the parameter ε controls the degree of generalization. If

ε is ø (the empty set), then R3 is the maximal consistent generalization of R1. If ε isD(x i) \ (A ∪ B) (where D(xi) is the domain of xi), then R3 is the minimal consistentgeneralization of R1 involving only xi. In AQ programs, the extension-against operator istypically used with ε = ø.

By repeating the extension-against operator until the resulting rule no longer covers anynegative examples, a consistent concept description (one that covers no negative examples)can be generated. Such a process can be applied in order to generate a description (cover)that is complete and consistent with regard to all the training examples.

By applying the extension-against operator with different values of the parameter ε, one cangenerate descriptions with different degrees of generality. For instance, in AQ15c, in orderto learn a characteristic rule, the output of the operator with ε initially set to ø is maximally

Page 25: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 3

specialized in such a way that it continues to cover all of the positive examples described bythe initial extension. If discriminant rules are desired, the extension will be maximallygeneralized so long as it continues not to cover any negative examples of the concept.

In order to effectively apply the extension-against operator to structured attributes, newgeneralization rules need to be defined. Let us illustrate the problem by an example thatuses a structured attribute “Food” shown in Figure 10. Each non-leaf node denotes aconcept that is more general than its children nodes. These relationships need to be takeninto consideration when generalizing given facts. Suppose that the concept to be learned isexemplified by statements: “John eats strip steak” and “John doesn’t eat vanilla ice cream.”There are many consistent generalizations of these facts, for example, “John eats stripsteak,” “John eats steak,” “John eats cattle,” “John eats meat,” “John eats meat orvegetables,” or “John eats anything but vanilla ice cream.” The first statement representsthe maximally specific description (no generalization), the last statement represents themaximally general description, and the remaining ones represent intermediate levels ofgeneralization. A problem arises in determining the generalization of most interest for agiven situation. We approach this problem by drawing insights from human reasoning.

Food

Meat Vegetable Dessert

Cattle Pigs Fowl Carrots Broccoli Beans Frozen Pies Pudding

Hamburger Steak Veal Green Pinto Baked Ice Cream Sherbet Cherry Apple

T-Bone Strip Vanilla Rocky Road+ -+

Anchor nodes are shown in bold. Nodes marked by + and — are valuesoccurring in positive and negative examples, respectively.

Figure 10. The domain of a structured attribute “Food.”

Cognitive scientists have noticed that people prefer certain nodes in a generalizationhierarchy (concepts) over other nodes when creating descriptions (e.g., Rosch et al, 1976).Factors that influence the choice of a concept (node) include the concept typicality (howcommon are a concept’s features among its sibling concepts), and the context in which theconcept is being used. For instance, upon seeing a robin (a typical bird), we may say,“There is a bird,” rather than “There is a robin,” assuming that the given situation does notrequire a specification of the type of bird. On the other hand, when we see a penguin, amuch less typical bird, we are more likely to say “There is a penguin,” rather than “There isa bird”. This way a listener (who is not an observer) will not assign to the unseen birdcharacteristics typical to a bird, but rather the special characteristics of a penguin. Thisfacilitates communication. Context also comes into play; at a gathering of bird watchers,the robin will not likely be called simply a bird, but rather will be referred to by itstaxonomic name.

Page 26: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 4

To provide some mechanism for capturing such preferences, INLEN allows a user todefine anchor nodes in a generalization hierarchy. Such nodes should reflect the interestsof a given application (Kaufman and Michalski, 1996). To illustrate this idea, considerFigure 10 again. In this hierarchy, vanilla and rocky road are kinds of ice cream; ice creamis a frozen dessert, which is a dessert, which is a type of food. In everyday usage,depending on the context, we will typically describe vanilla or rocky road as ice cream ordessert, but less likely as frozen dessert or food. Hence, we can designate dessert and icecream as anchor nodes in the Food hierarchy. Using information about anchor nodes,different rule preference criteria can be specified, such as selecting the rule with the mostgeneral anchor nodes, or the one that generalizes positive examples to the next higheranchor node(s).

INLEN supports the use of structured attributes both as independent (input) and dependent(output) variables. Structured independent attributes represent hierarchies of values that areused to characterize entities. Structured dependent attributes represent hierarchies ofdecisions or classifications that can be made about an entity. Through the use of structuredoutput attributes, INLEN’s learning module can determine rules at different levels ofgenerality.

While dependent attributes, like independent ones, can take on in principle different types(nominal, linear, cyclic or structured), in practical applications they are frequently eithernominal or linear. A nominal output attribute is most frequently used in concept learning;its values denote concepts or classes to be learned. A linear output attribute (which istypically a measurement on a ratio scale) is used to denote a measurement whose values areto be predicted on the basis of the past data.

In many applications, it is desirable to use a structured attribute as a dependent variable.For example, when deciding which personal computer to buy, one may first decide thegeneral type of the computer—whether it is to be IBM PC-compatible or Macintosh-compatible. After deciding the type, one can focus on a specific model of the chosen type.The above two-level decision process is easier to execute than a one-level process in whichone has to directly decide which computer to select from a large set.

When a dependent variable is structured, the learning operator focuses first on the top-levelvalues (nodes), and creates rules for them. Subsequently, it creates rules for thedescendant nodes in the context of their ancestors. This procedure produces decision rulesthat are simpler and easier to interpret than rules learned with a flat (nominal) organizationof the decision attribute.

8 Learning Decision Structures from Decision Rules

One of the main reasons for data exploration is to learn rules or patterns in data that willenable a data analyst to predict future cases. Thus, when such rules are learned, one needsa method for efficiently applying the rules for prediction. Since a convenient structure forimplementing a decision process is a decision tree, the problem of how to transferknowledge to a decision tree arises. In the conventional machine learning approach,decision trees are learned directly from training examples, thus avoiding the step of firstcreating rules (Hunt, Marin and Stone, 1966; Quinlan, 1986; 1993).

Learning a decision tree directly from examples, however, may have serious disadvantagesin practice. A decision tree is a form of procedural knowledge. Once it has beenconstructed, it is not easy to modify it to accommodate changes in the decision-making

Page 27: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 5

conditions. For example, if an attribute (test) assigned to a high-level node in the tree isimpossible or too costly to measure, the decision tree offers no alternative course of actionother than probabilistic reasoning (Quinlan, 1986).

In contrast, a human making the decision would probably search for alternative tests toperform. People can do this because they typically store decision knowledge in adeclarative form. From a declarative form of knowledge, such as a set of decision rules,one can usually construct many different, but logically equivalent, or nearly equivalent,decision trees. One such decision tree may be preferable to another in a given decision-making situation. Therefore, it is desirable to store knowledge decoratively and to transferit only when the need arises to the procedural form that is most appropriate to the givensituation.

Another weakness of decision trees is that they may become unwieldy andincomprehensible because of their limited knowledge representational power. To overcomethe above limitations, a new approach has been developed that creates task-orienteddecision structures from decision rules (Imam, 1995, Michalski and Imam, 1997). Adecision structure is a generalization of a decision tree in which tests associated with nodescan refer not only to single attributes, but also to functions of multiple attributes; branchesmay be associated not only with single values/results of these tests, but also with a set ofsuch values; and leaves can be assigned not only a single decision, but also a set ofalternative decisions with appropriate probabilities.

This approach has been implemented in the AQDT-2 program, and employs an AQ-typelearning algorithm (AQ15c and AQ17-DCI) for determining decision rules from examples.Among its advantages are the ability to generate a decision structure that is most suitable toa particular task and the ability to avoid or delay measuring costly attributes. Differentusers may want to generate different decision structures from a given set of rules, so thatthe structures are tailored to their individual situations. Furthermore, if an attribute isdifficult to measure, or cannot be measured at all, the program can be instructed to build adecision structure from rules that tries to avoid this attribute, or measure it only whennecessary.

Another advantage of this methodology is that once a rule set is determined, a decisionstructure can be generated from it far more rapidly than if it has to be determined fromexamples, hence processing time is very small. Also, a set of rules will take up lessstorage space than the data set from which it was learned.

Experiments with AQDT-2 indicate that decision structures learned from decision rules tendto be significantly simpler than decision trees learned from the same data, and frequentlyalso have a higher predictive accuracy. For example, a decision structure learned byAQDT-2 for a wind bracing design problem had 5 nodes and 9 leaves, with a predictiveaccuracy of 88.7% when tested against a new set of data, while the decision tree generatedby the popular program C4.5 had 17 nodes and 47 leaves with a predictive accuracy of84% (Michalski and Imam, 1997). In another experiment, a decision tree learned fromdecision rules by AQDT to analyze Congressional voting patterns had 7 nodes and 13leaves, with a predictive accuracy of 91.8% (when AQDT built an equivalent decisionstructure by combining some branches, the number of leaves was reduced to 8), while thedecision tree learned by C4.5 from the same set of training examples had 8 nodes and 15leaves, with a predictive accuracy of 85.7% (Imam and Michalski, 1993).

Page 28: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 6

This methodology directly fits the philosophy of INLEN. A rule base may be providedeither from an expert or through the use of a rule learning operator, thereby allowing for thegeneration of decision structures from rules.

9 Automatic Improvement of Representation Spaces9 . 1 Determining Most Relevant Attributes

In a large database, many attributes may be used to characterize given entities. For anyspecific problem of determining rules characterizing the relationship between a designatedoutput attribute and other attributes, it may be desirable to limit the independent attributes tothe most relevant ones. To this end, one may use many different criteria for evaluating therelevance of an attribute for a given classification problem, such as gain ratio (Quinlan,1993), gini index (Breiman et al, 1984), PROMISE (Baim, 1982), and chi-square analysis(Hart, 1984; Mingers, 1989).

These criteria evaluate attributes on the basis of their expected global performance, whichmeans that those attributes with the highest ability to discriminate among all classes areselected as the most relevant.

When determining a declarative knowledge representation, such as decision rules, the goalis somewhat different. Here, each class is described independently from other classes, andthe simplest and most accurate rules for each class are desired. Hence, if an attribute has asingle value that characterizes very well just one specific class, the attribute with this valuewill be used effectively in a corresponding decision rule. In contrast, such an attribute mayhave a low global discriminating value, and is thus ignored in building a decision tree. Itfollows that the determination of attributes for decision trees and for decision rules needs tofollow different criteria.

To illustrate this point, consider the problem of recognizing the upper-case letters of theEnglish alphabet. Two of the attributes to be considered might be whether the letter has atail and whether it is made up exclusively of straight lines. In a rule-based (declarative)representation, the letter Q can be distinguished from the rest of the alphabet by a simpleand concise property, if the letter has a tail, it is a Q. Conversely, the straight line conditionis alone insufficient to discriminate any specific letter, but is useful overall.

Thus, the attribute has-tail is very useful for learning one specific class, although not veryuseful for characterizing other classes. It is thus appropriate for use in rule learning. Indecision-tree learning, however, it may be evaluated as having a relatively low overallutility and replaced by other attributes. This will most likely happen if Qs are relativelyrare. Hence, testing the letter for a tail will be considered a wasted operation, as it onlyserves to eliminate the possibility of it being a Q, without making any progress indistinguishing between the other 25 letters. Meanwhile, testing the condition all-straight-lines immediately bisects the search space. It is better to pare down the set of hypothesesmore rapidly, and only check for a tail as a last step when the set of possible letters hasbeen reduced to O and Q. This way, the recognition of Q will require more tests thannecessary, but at no expense to the recognition of other letters.

INLEN supports both global and local attribute evaluation criteria for selecting the mostrelevant attributes. The former is based on the PROMISE methodology (Baim, 1982),while the latter employs a variation of PROMISE that is oriented toward the maximumperformance of some attribute value, rather than on the attribute’s global performance.

Page 29: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 7

9 . 2 Generating New Attributes

When the original representation space is weakly relevant to the problem at hand, or theconcept to be learned is difficult to express in the from of attributional decision rules suchas those employed in INLEN, there is a need to generate new attributes that are functions ofthe original ones and better suited to the given problem. This is done by a constructiveinduction operator based on the program AQ17-DCI (Bloedorn and Michalski, 1996).

In the case of a database that contains information on objects changing over time, one needsa mechanism for constructive induction that can take advantage of the time data ordering.For example, the database may contain information on the maximum temperature at a givenlocation each day, with a field in each record indicating the day on which its temperaturewas recorded. Inherent in a timestamped representation are many attributes that can begenerated through constructive induction, for example, date of the highest temperature, theminimum population growth rate during some period, weediness on date of planting, etc.

CONVART (Davis, 1981) uses user-provided and default system suggestions to search foruseful time-dependent attributes that are added to the representation space. It uses the itemson the suggestion list to generate new attributes and to test them for likely relevance to theproblem. If they exceed a relevance threshold, it adds them to the representation space,repeating this procedure until a desired number of new attributes have been constructed.As part of its attribute construction capability, INLEN will incorporate such techniques forthe generation of time-dependent attributes.

1 0 Exemplary Application: Discovery in Economic andDemographic Data

1 0 . 1 Motivation

Economic analysis is one domain in which conceptual data exploration tools can be of greatvalue. The following example illustrates the role an intelligent data exploration system canplay in the extraction of knowledge from data.

The United States government maintains records of the import and export of goodsfrom various countries of the world. The different products and raw materials aredivided and subdivided into different categories. In the early 1980s the data showed asharp decline in the import of trucks from Japan and a corresponding increase in theimport of auto parts from Japan. It took several years before analysts noticed that factand concluded that Japan was shipping the chassis and truck beds separately to theUS, where they would be subsequently assembled, thereby avoiding a high US tariffon imported trucks that was directed primarily at Europe and had been on the bookssince World War II. When United States analysts inferred this explanation, the USand Japan commenced trade negotiations pertaining to the import of trucks.

How much sooner would that trend have been noticed had a conceptual data explorationprogram been applied to the data and pointed out the opposite changes in two relatedcategories to an analyst? How much revenue did the undiscovered truth cost the US beforethey could finally work out a new agreement with Japan? Noticing economic trends andpatterns like the one above is a difficult task, as humans can easily get overwhelmed by theamount of data.

Page 30: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 8

Based on such motivation, the analysis of economic and demographic data has become oneof the focus domains for INLEN development and testing. We illustrate some of itsdiscovery capabilities through experiments involving two similar data sets: one providedby the World Bank consisting of information on 171 countries for the period of 1965 to1990 (in terms of 95 attributes), and one extracted from the 1993 World Factbook(published by the Central Intelligence Agency) containing several databases of informationon 190 countries (in terms of 17 attributes).

1 0 . 2 Experiment 1: Integration of Multiple Operators

The World Bank data enabled us to conduct a number of experiments for testing INLENcapabilities. One experiment focused on distinguishing between development patterns inEastern Europe and East Asia, first by identifying such patterns, and then by generatingdiscriminant rules (Kaufman, 1994).

A conceptual clustering operator determined a way of grouping the countries, based oneach country’s change in the percentage of its population in the labor force between 1980and 1990. In this classification, the typical Eastern European country and the typical EastAsian country fell into separate groups. Most of the European countries had a labor forcechange below a threshold determined for the region by the clustering program, while mostof the Asian countries had changes in labor force participation above the thresholddetermined for their region.

Based on this grouping, the rule learning operator (using the AQ15c inductive learningprogram) was called upon first in characteristic mode to characterize the Asian-likecountries (those above their regional thresholds) and the European-like countries (thosebelow their regional thresholds), and then in discriminant rule-optimizing mode tocondense those characterizations into simple discriminant rules. The discriminant rulesobtained were:

Country is Asian-Like if:A.1 Change in Labor Force Participation ≥ slight_gain, (9 countries)

orB.1 Life Expectancy is in 60s, and 2 Working Age Population ≤ 64%, (2 countries)

Country is European-Like if:A.1 Change in Labor Force Participation is near 0 or decreasing, and 2 Life Expectancy is not in 60s, (7 countries)

orB.1 Percentage of Labor Force in Industry ≥ 40. (1 country)

The rules show that of the 10 attributes in the original data set, only four attributes areinstrumental in distinguishing between the European-style and Asian-style developmentpatterns, namely Change in Labor Force Participation, Life Expectancy, Working AgePopulation and Percentage of Labor Force in Industry. In both the Asian- and European-Like cases, the first rule accounted for most of the countries fitting the class, while thesecond one described the remainder.

This experiment demonstrated one of the cornerstone features of the methodology - anintegration of different learning and discovery strategies that allows knowledge to bepassed from one operator to another in a seamless way, leading to conclusions unreachable

Page 31: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

2 9

by any one individual program. It also shows that the rules created by the system are easyto understand and interpret.

1 0 . 3 Experiment 2: Detecting Anomalies in Subgroups

Another experiment with INLEN investigated the problem of detecting interestingregularities within the subgroups it creates. While the subgroups in a demographic domainmay indicate that member countries or regions have something in common, notableexceptions may be exposed when a member of these constructed subsets shows a markeddissimilarity to the rest of the group. These exceptions in turn may prove to be aspringboard for further discovery.

INLEN discovered several rules from the World Factbook PEOPLE database characterizingthe 55 countries with low (less than 1% per year) population growth rates by invoking therule learning operator in characteristic mode. One of the characteristic descriptions (Figure11) had three conditions that together characterized 19 low growth countries and only onewith higher population growth rates.

Characteristic Description of Countries with Population Growth Ratebelow 1 per 1000 people: Pos Neg Supp Comm

1 Birth Rate = 10 to 20 or Birth Rate ≥ 50 46 20 69% 84%2 Predominant Religion is Orthodox or 40 68 37% 73%

Protestant or Hindu or Shinto3 Net Migration Rate ≤ +20 32 104 23% 58%

All 3 conditions: 19 1 95% 35%

Figure 11. A characterization of countries with low population growth.

In the characterization shown in Figure 11, the columns Pos and Neg respectively representthe number of positive and negative examples satisfying the condition. The support level(Supp) is defined as Pos / (Pos + Neg), giving an indication of how much support thecondition lends to the suggestion that a country’s Population Growth Rate is less than 1%.The commonality level (Comm) is defined as Pos / Total_Pos, giving an indication of howcommonly the condition occurs in countries with Population Growth Rates below 1% (inthis example, Total_Pos = 55).

The first condition (and thus the strongest in terms of support level) states that countrieswith population growth rate below 1% have a low (under 20 per 1000 population) or veryhigh (over 50 per 1000 population) birth rate. The presence of a very high birth rate incountries with low population growth is highly counterintuitive; examination of the 19countries covered by the description pointed out that 18 had birth rates below 20, whileonly one, Malawi, had the high birth rate. When further attention was focused on Malawi,the explanation was clear. Malawi had a massive outward net migration rate of over 30 per1000 population, by far the most extreme migration rate in the world. INLEN thusfacilitated a discovery of a surprising exception to a normal pattern.

1 0 . 4 Experiment 3: Utilizing Structured Attributes

The rule shown in the previous example contained an attribute “predominant religion.”This attribute was presented as a nominal attribute in the initial dataset. In order to examinehow the structuring of attributes affects knowledge discovery, INLEN was applied to

Page 32: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3 0

identical data sets with and without the Religion attribute being structured (Kaufman andMichalski, 1996). A portion of the attribute domain structure is shown in Figure 12.

One strong argument for structuring is that if the Predominant Religion attribute has beenset up in an unstructured (nominal) manner, the statement “Predominant Religion isLutheran” would be regarded as being as antithetical to “Predominant Religion is Christian”as it is to the statement “Predominant Religion is Buddhist,” since “Lutheran,” “Christian”and “Buddhist” are all considered equally different in a “flat” domain. This would lead tothe possibility that some contradictions such as “Predominant Religion is Lutheran, but notChristian” might be generated.

Predominant Religion

Muslim Jewish Buddhist Shinto Christian Hindu

Sunni Shi’a Ibadhi Theravada Protestant Orthodox Roman Catholic

Lutheran Evangelical Georgian Bulgarian Romanian Anglican Tuvalu Armenian Greek

Figure 12. Part of the structure of the PEOPLE database’s Religion attribute.

Experiments using INLEN-2 have lent support to this and other hypotheses regarding theuse of structured and non-structured attributes. Among the findings regarding their use asindependent variables was that structuring attributes leads to simpler rules than when notstructuring them. For example, when INLEN learned rules to distinguish the 55 countrieswith low population growth rate (less than 1%) from other countries, in a version of thePEOPLE database in which the attribute “Predominant Religion” was not structured, one ofthe rules it found was:

Population Growth Rate < 1% if: (20 examples)1 Literacy = 95% to 99%,2 Life Expectancy is 70 to 80 years,3 Predominant Religion is Roman Catholic or Orthodox or Romanian or Lutheran or

Evangelical or Anglican or Shinto,4 Net Migration Rate ≤ +20 per 1000 population.

This rule was satisfied by 20 of the 55 countries with low growth rates. When the sameexperiment was run with “Religion” used as a structured attribute, a simpler pattern wasdiscovered:

Population Growth Rate < 1% if: (21 examples, 1 exception)1 Literacy = 95% to 99%,2 Life Expectancy is 70 to 80 years,3 Predominant Religion is Christian or Shinto,4 Net Migration Rate ≤ +10 per 1000 population.

Page 33: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3 1

This rule has one exception (the United States, whose 1993 population growth rate wasbetween 1% and 2%). If full consistency is required, the third condition could still beexpressed in a simpler form than in an unstructured religion domain by performing aminimal specialization operation on the node Christian so that the rule would cover thesame positive examples, but not the exception.

Similar differences were obtained by structuring dependent attributes. By arranging eventsinto different levels of generality, rules classified them accordingly, which reduced thecomplexity and increased the informational significance of the rules at different levels ofgeneralization.

These effects were especially visible at the lower levels of the hierarchy. In theunstructured dataset, five rules, each with two to five conditions, were required to definethe 11 Sunni Muslim countries. The only one to describe more than two of the 11countries was a rule with quite fragmented conditions:

Predominant Religion is Sunni_Muslim if: (4 examples)1 Literacy ≠ 30% to 99%,2 Infant Mortality Rate is 25 to 40 or greater than 55 per 1000 population3 Fertility Rate is 1 to 2 or 4 to 5 or 6 to 7 per 1000 population,4 Population Growth Rate is 1% to 3% or greater than 4%.

The value ranges in these conditions are divided into multiple segments, suggesting thatthis is not a strong pattern. In contrast, using a structured religion attribute, the learningoperator produced two simple and easily understood patterns, each with one onlycondition:

Predominant Religion is Sunni_Muslim if: (10 examples, 1 exception)1 Infant Mortality Rate ≥ 40 per 1000 population.

Predominant Religion is Sunni_Muslim if: (4 examples)1 Birth Rate is 30 to 40 per 1000 population.

As described above, these rules apply only in the context of predominantly Islamiccountries, and are based on the assumption that that determination has already been made.

1 0 . 5 Experiment 4: Applying Constructive Induction Operators

An experiment chronicled by Bloedorn and Michalski (1996) demonstrates the power ofutilizing constructive induction as a knowledge discovery operator. Working from 11economic attributes sampled over each of five consecutive years, 1986-1990 (for a total of55 available attributes per record), the learning program attempted to discover rules topredict countries’ changes in gross national product over the 5-year period. By applyingthree data-driven constructive induction operators—generating new attributes based on theexisting attribute set, removing attributes less relevant to the goal concept, and abstractingnumerical attributes into a small number of intervals—the predictive accuracy on new dataincreased by nearly half (from 41.7% to 60.5%).

Among the newly constructed highly relevant attributes were Change in EnergyConsumption Between 1986 and 1988, Ratio of Birth Rate in 1989 to EnergyConsumption in 1990, and Average Annual Energy Consumption Over the 5-year Period.

Page 34: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3 2

These results demonstrated that constructive induction can be a very useful tool foranalyzing data, as it can build more adequate representation spaces for knowledgediscovery.

1 1 Summary

The main thesis of this paper is that modern methods developed in symbolic machinelearning have a direct and important application to the development of new operators forconceptual data exploration. A wide range of ideas on the applicability of various machinelearning methods to this area were presented.

Two highly important operators are the construction of conceptual hierarchies (conceptualclustering), and the inductive derivation of general rules characterizing the relationshipbetween designated output and input attributes. These rules represent high-levelknowledge that can be of great value to a data analyst and directly usable in humandecision-making. Other important operators include construction of equations along withlogical preconditions for their application, determination of symbolic descriptions of timesequences, selection of most relevant attributes, generation of new, more relevantattributes, and selection of representative examples.

In contrast to many data mining approaches, the presented methodology requires aconsiderable amount of background knowledge regarding the data and the domain ofdiscourse. This background knowledge may include, for example, a specification of thedomain and the type of the attributes, the relationships among them, causal dependencies,theories about the objects or processes that generated the data, goals of the data analysisand other high-level knowledge. An important aspect of the methodology is its ability totake advantage of this knowledge.

The machine learning techniques implemented in the INLEN system allow a user toperform easily a wide range of symbolic data manipulation and knowledge generationoperations. The illustrative examples demonstrate a significant potential utility of thedescribed multistrategy methodology in solving problems of data mining and knowledgediscovery.

References

Baim, P.W., “The PROMISE Method for Selecting Most Relevant Attributes for InductiveLearning Systems,” Report No. UIUCDCS-F-82-898, Department of Computer Science,University of Illinois, Urbana, 1982.

Bentrup, J.A., Mehler, G.J. and Riedesel, J.D., “INDUCE 4: A Program forIncrementally Learning Structural Descriptions From Examples,” Reports of the IntelligentSystems Group, ISG 87-2. UIUCDCS-F-87-958, Department of Computer Science,University of Illinois, Urbana, 1987.

Bergadano, F., Matwin, S., Michalski, R.S. and Zhang, J., “Learning Two-TieredDescriptions of Flexible Concepts: The POSEIDON System,” Machine Learning, 8, pp. 5-43, 1992.

Page 35: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3 3

Bloedorn, E. and Michalski, R.S., “The AQ17-DCI System for Data-Driven ConstructiveInduction and Its Application to the Analysis of World Economics,” Proceedings of the 9thInternational Symposium on Methodologies for Intelligent Systems, Zakopane, Poland,1996.

Bloedorn, E., Wnek, J. and Michalski, R.S., “Multistrategy Constructive Induction,”Proceedings of the Second International Workshop on Multistrategy Learning, HarpersFerry, WV, pp. 188-203, 1993.

Bongard, N., Pattern Recognition, Spartan Books, New York (a translation fromRussian), 1970.

Bratko, I., Muggleton, S. and Karalic, A., “Applications of Inductive LogicProgramming,” In Michalski, R.S., Bratko, I. and Kubat, M. (eds.), Machine Learningand Data Mining: Methods and Applications, London, John Wiley & Sons, 1997.

Breiman, L., Friedman, J.H., Olshen, R.A. and Stone, C.J., Classification andRegression Trees, Belmont, CA, Wadsworth Int. Group, 1984.

Carpineto, C. and Romano, G., “Some Results on Lattice-based Discovery in Databases,”Workshop on Statistics, Machine Learning and Knowledge Discovery in Databases,Heraklion, pp. 216-221, 1995a.

Carpineto, C. and Romano, G., “Automatic Construction of Navigable Concept NetworksCharacterizing Text Databases,” Gori, M. and Soda, G. (eds.), Topics in ArtificialIntelligence, LNAI 992-Springer, pp. 67-78, 1995b.

Collins, A. and Michalski, R.S., “Toward a Formal Theory of Human PlausibleReasoning,” Proceedings of the Third Annual Conference of the Cognitive ScienceSociety, Berkeley, CA, 1981.

Collins, A. and Michalski, R.S., “The Logic of Plausible Reasoning: A Core Theory,”Cognitive Science, 13, pp. 1-49, 1989.

Daniel, C. and Wood, F.S., Fitting Equations to Data, New York, John Wiley & Sons,1980.

Davis, J., “CONVART: A Program for Constructive Induction on Time-Dependent Data,”M.S. Thesis, Department of Computer Science, University of Illinois, Urbana, 1981.

Diday, E. (ed.), Proceedings of the Conference on Data Analysis, Learning Symbolic andNumeric Knowledge, Nova Science Publishers, Inc., Antibes, 1989.

Dieterrich, T. and Michalski, R.S., “Learning to Predict Sequences,” Chapter in MachineLearning: An Artificial Intelligence Approach Vol. 2. Michalski, R.S., Carbonell, J.G. andMitchell, T.M. (eds.), Morgan Kaufmann, pp. 63-106, 1986.

Dontas, K., “APPLAUSE: An Implementation of the Collins-Michalski Theory ofPlausible Reasoning,” M.S. Thesis, Computer Science Department, The University ofTennessee, Knoxville, TN, 1988.

Dubois, D., Prade, H. and Yager, R.R. (eds.), Readings in Fuzzy Sets and IntelligentSystems, Morgan Kaufmann, 1993.

Page 36: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3 4

Evangelos S. and Han, J. (eds.), Proceedings of the Second International Conference onKnowledge Discovery and Data Mining, Portland, OR, 1996.

Falkenhainer, B.C. and Michalski, R.S., “Integrating Quantitative and QualitativeDiscovery in the ABACUS System,” In Kodratoff, Y. and Michalski, R.S. (eds.),Machine Learning: An Artificial Intelligence Approach Vol. III , San Mateo, CA, MorganKaufmann, pp. 153-190, 1990.

Fayyad, U., Piatetsky-Shapiro, G. and Smyth, P., “Knowledge Discovery and DataMining: Toward a Unifying Framework,” Proceedings of the Second InternationalConference on Knowledge Discovery and Data Mining. Portland, OR, pp. 82-88, 1996.

Fayyad, U.M. Piatetsky-Shapiro, G. Smyth, P. and Uhturusamy, R. (eds.), Advances inKnowledge Discovery and Data Mining, San Mateo, CA, AAAI Press, 1996.

Feelders, A. Learning from Biased Data Using Mixture Models,” Proceedings of theSecond International Conference on Knowledge Discovery and Data Mining, Portland,OR, pp. 102-107, 1996.

Forsyth, R. and Rada, R., Machine Learning: Applications in Expert Systems andInformation Retrieval, Pittman, 1986.

Greene, G., “The Abacus.2 System for Quantitative Discovery: Using Dependencies toDiscover Non-Linear Terms,” Reports of the Machine Learning and Inference Laboratory,MLI 88-4, Machine Learning and Inference Laboratory, George Mason University,Fairfax, VA, 1988.

Hart, A., “Experience in the Use of an Inductive System in Knowledge Engineering,” In MBramer (Ed.), Research and Developments in Expert Systems, Cambridge, CambridgeUniversity Press, 1984.

Hong, J., Mozetic, I. and Michalski, R.S., “AQ15: Incremental Learning of Attribute-Based Descriptions from Examples: The Method and User’s Guide,” Reports of theIntelligent Systems Group, ISG 86-5, UIUCDCS-F-86-949, Department of ComputerScience, University of Illinois, Urbana, 1986.

Hunt, E., Marin, J. and Stone, P., Experiments in Induction, Academic Press, New York,1966.

Imam, I.F., “Discovering Task-Oriented Decision Structures from Decision Rules,” Ph.D.dissertation, School of Information Technology and Engineering, George MasonUniversity, Fairfax, VA, 1995.

Imam, I.F. and Michalski, R.S., “Should Decision Trees be Learned from Examples orFrom Decision Rules?” Proceedings of the Seventh International Symposium onMethodologies for Intelligent Systems (ISMIS-93), Trondheim, Norway, 1993.

Kaufman, K., “Comparing International Development Patterns Using Multi-OperatorLearning and Discovery Tools,” Proceedings of the AAAI-94 Workshop on KnowledgeDiscovery in Databases, Seattle, WA, pp. 431-440, 1994.

Page 37: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3 5

Kaufman, K.A. and Michalski, R.S., “EMERALD: An Integrated System of MachineLearning and Discovery Programs to Support Education and Experimental Research,”Reports of the Machine Learning and Inference Laboratory, MLI 93-10, Machine Learningand Inference Laboratory, George Mason University, Fairfax, VA, 1993.

Kaufman, K. and Michalski, R.S., “A Method for Reasoning with Structured andContinuous Attributes in the INLEN-2 Knowledge Discovery System,” Proceedings of the2nd International Conference on Knowledge Discovery and Data Mining, Portland, OR,pp. 232-237, 1996.

Kaufman, K., Michalski, R.S. and Kerschberg, L., “Mining for Knowledge in Data:Goals and General Description of the INLEN System,” In Piatetsky-Shapiro, G. andFrawley, W.J. (Eds.), Knowledge Discovery in Databases, Menlo Park, CA: AAAI Press,pp. 449-462, 1991.

Kerber, R., “ChiMerge: Discretization of Numeric Attributes. Proceedings of the TenthNational Conference on Artificial Intelligence (AAAI-92), San Jose, CA, pp. 123-127,1992.

Kodratoff, Y, Introduction to Machine Learning, Pittman, 1988.

Kokar, M.M., “Coper: A Methodology for Learning Invariant Functional Descriptions,”Chapter in Machine Learning: A Guide to Current Research, Michalski, R.S., Mitchell,T.M and Carbonell, J.G. (Eds.), Kluwer Academic Publishers, Boston, MA 1986.

Kodratoff, Y. and Michalski, R.S. (eds.), Machine Learning: An Artificial IntelligenceApproach, Vol. III. San Mateo, CA, Morgan Kaufmann., 1990

Lakshminarayan, K., Harp, S.A., Goldman, R. and Samad, T., “Imputation of MissingData Using Machine Learning Techniques,” Proceedings of the Second InternationalConference on Knowledge Discovery and Data Mining, Portland, OR, pp. 140-145, 1996.

Langley, P., Bradshaw G.L. and Simon, H.A., “Rediscovering Chemistry with theBACON System,” Machine Learning: An Artificial Intelligence Approach, Michalski,R.S., Carbonell, J.G. and Mitchell, T.M. (Eds.), Morgan Kaufmann, San Mateo, CA, pp.307-329, 1983.

Larson, J.B, “INDUCE-1: An Interactive Inductive Inference Program in VL21 LogicSystem,” Report No. 876, Department of Computer Science, University of Illinois,Urbana, 1977.

Maloof, M.A. and Michalski, R.S, “Learning Evolving Concepts Using Partial MemoryApproach,” Working Notes of the 1995 AAAI Fall Symposium on Active Learning,Boston, MA, pp. 70-73, 1995.

Michael, J., “Validation, Verification and Experimentation with Abacus2,” Reports of theMachine Learning and Inference Laboratory, MLI 91-8, Machine Learning and InferenceLaboratory, George Mason University, Fairfax, VA, 1991.

Michalski, R.S., “A Planar Geometrical Model for Representing Multi-DimensionalDiscrete Spaces and Multiple-Valued Logic Functions,” ISG Report No. 897, Departmentof Computer Science, University of Illinois, Urbana, 1978.

Page 38: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3 6

Michalski, R.S., “Inductive Learning as Rule-Guided Generalization and ConceptualSimplification of Symbolic Descriptions: Unifying Principles and a Methodology,”Workshop on Current Developments in Machine Learning, Carnegie Mellon University,Pittsburgh, PA, 1980.

Michalski, R.S., “A Theory and Methodology of Inductive Learning,” In Michalski, R.S.Carbonell, J.G. and Mitchell, T.M. (eds.), Machine Learning: An Artificial IntelligenceApproach, Palo Alto: Tioga Publishing, pp. 83-129, 1983.

Michalski, R.S, “Learning Flexible Concepts: Fundamental Ideas and a Method Based onTwo-tiered Representation,” In Kodratoff, Y. and Michalski, R.S. (eds.), MachineLearning: An Artificial Intelligence Approach, Vol. III , San Mateo, CA, MorganKaufmann, pp. 63-102, 1990.

Michalski, R.S., “Searching for Knowledge in a World Flooded with Facts,” in AppliedStochastic Models and Data Analysis, Vol. 7, pp. 153-l63, January, 1991.

Michalski, R.S., “Inferential Theory of Learning: Developing Foundations forMultistrategy Learning,” In Michalski, R.S. and Tecuci, G. (eds.), Machine Learning: AMultistrategy Approach, San Francisco: Morgan Kaufmann, pp. 3-61, 1994.

Michalski, R.S., Baskin, A.B. and Spackman, K.A., “A Logic-based Approach toConceptual Database Analysis,” Sixth Annual Symposium on Computer Applications inMedical Care (SCAMC-6), George Washington University, Medical Center, Washington,DC, pp. 792-796, 1982.

Michalski, R.S., Carbonell, J.G. and Mitchell, T.M. (eds.), Machine Learning: AnArtificial Intelligence Approach, Palo Alto, CA, Tioga Publishing, 1983.

Michalski, R.S., Carbonell, J.G. and Mitchell, T.M. (eds.), Machine Learning: AnArtificial Intelligence Approach Vol. 2, San Mateo, CA, Morgan Kaufmann, 1986.

Michalski, R.S. and Imam, I.F., “On Learning Decision Structures,” FundamentaMatematicae, Polish Academy of Sciences, 1997 (in press).

Michalski, R.S., Kaufman, K. and Wnek, J., “The AQ Family of Learning Programs: AReview of Recent Developments and an Exemplary Application,” Reports of the MachineLearning and Inference Laboratory, MLI 91-11, Machine Learning and InferenceLaboratory, George Mason University, Fairfax, VA, 1991.

Michalski, R.S., Kerschberg, L., Kaufman, K. and Ribeiro, J., “Mining for Knowledgein Databases: The INLEN Architecture, Initial Implementation and First Results,” Journalof Intelligent Information Systems: Integrating AI and Database Technologies, 1, pp. 85-113, August 1992.

Michalski, R. S., Ko, H. and Chen, K., “SPARC/E(V.2), An Eleusis Rule Generator andGame Player”, Reports of the Intelligent Systems Group, ISG No. 85-11, UIUCDCS-F-85-941, Department of Computer Science, University of Illinois, Urbana, 1985.

Michalski, R.S., Ko, H. and Chen, K., “Qualitative Prediction: A Method and a ProgramSPARC/G,” Chapter in Expert Systems, Guetler, C. (ed.), Academic Press, Inc., London,1986.

Page 39: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3 7

Michalski, R.S. and Larson, J.B., “Selection of Most Representative Training Examplesand Incremental Generation of VL1 Hypotheses: The Underlying Methodology and theDescription of Programs ESEL and AQ11,” Report No. 867, Department of ComputerScience, University of Illinois, Urbana, 1978.

Michalski, R.S. and McCormick, B.H., “Interval Generalization of Switching Theory,”Proceedings of the 3rd Annual Houston Conference on Computer and System Science,Houston, TX, 1971.

Michalski, R.S. , Mozetic, I., Hong, J. and Lavrac, N., “The AQ15 Inductive LearningSystem: An Overview and Experiments,” ISG Report 86-20, UIUCDCS-R-86-1260,Department of Computer Science, University of Illinois, Urbana, 1986.

Michalski, R.S., Stepp, R. and Diday, E., “A Recent Advance in Data Analysis: ClusteringObjects into Classes Characterized by Conjunctive Concepts,” Chapter in Kanal, L. andRosenfeld, A. (eds.), Progress in Pattern Recognition, Vol. 1, pp. 33-55, 1981.

Mingers, J., “An Empirical Comparison of Selection Measures for Decision-TreeInduction,” Machine Learning, 3, pp. 319-342, 1989.

Morgenthaler, S. and Tukey, J.W., “The Next Future of Data Analysis,” In Diday, E.(ed.), Proceedings of the Conference on Data Analysis, Learning Symbolic and NumericKnowledge, Nova Science Publishers, Inc., Antibes, 1989.

Pawlak, Z, Rough Sets: Theoretical Aspects of Reasoning about Data, Kluwer AcademicPublishers, Dordrecht, 1991.

Quinlan, J.R., “Induction of Decision Trees,” Machine Learning, 1, pp. 81-106, 1986.

Quinlan, J.R., “Probabilistic Decision Trees,” In Y. Kodratoff and R.S. Michalski, (Eds.),Machine Learning: An Artificial Intelligence Approach, Volume III , Morgan Kaufmann,San Mateo, CA, pp. 140-152, 1990.

Quinlan, J.R., “C4.5: Programs for Machine Learning,” Morgan Kaufmann, Los Altos,CA, 1993.

Reinke, R.E., “Knowledge Acquisition and Refinement Tools for the ADVISE Meta-Expert System,” Master’s Thesis, Department of Computer Science, University of Illinois,Urbana, 1984.

Reinke, R.E. and Michalski, R.S., “Incremental Learning of Concept Descriptions: AMethod and Experimental Results,” Machine Intelligence, 11, pp. 263-288, 1988.

Ribeiro, J., Kaufman, K. and Kerschberg, L., “Knowledge Discovery from MultipleDatabases,” Proceedings of the First International Conference on Knowledge Discoveryand Data Mining, Montreal PQ, pp. 240-245, 1995.

Rosch, E., Mervis, C., Gray, W., Johnson, D. and Boyes-Braem, P., “Basic Objects inNatural Categories,” Cognitive Psychology, 8, pp. 382-439, 1976.

Sharma, S., Applied Multivariate Techniques, London, John Wiley & Sons, Inc., 1996.

Page 40: Data Mining and Knowledge Discovery: A Review of …Data Mining and Knowledge Discovery: A Review of Issues and a Multistrategy Approach∗ Ryszard S. Michalski and Kenneth A. Kaufman

3 8

Slowinski, R. (ed.), Intelligent Decision Support: Handbook of Applications and Advancesof the Rough Sets Theory, Dordrecht/Boston/London, Kluwer Academic Publishers,1992.

Stepp, R., “A Description and User’s Guide for CLUSTER/2, A Program for ConjunctiveConceptual Clustering,” Report No. UIUCDCS-R-83-1084, Department of ComputerScience, University of Illinois, Urbana, 1984.

Tukey, J.W., The Collected Works of John W. Tukey, Vol. V, Philosophy and Principlesof Data Analysis: 1965-1986, Jones, L.V. (ed.), Wadsworth & Brooks/Cole, Monterey,CA, 1986.

Van Mechelen, I., Hampton, J., Michalski, R.S. and Theuns, P. (eds.), Categories andConcepts: Theoretical Views and Inductive Data Analysis, London, Academic Press, 1993.

Widmer, G. and Kubat, M., “Learning in the Presence of Concept Drift and HiddenConcepts,” Machine Learning, 23, pp. 69-101, 1996.

Wnek, J., “DIAV 2.0 User Manual, Specification and Guide through the DiagrammaticVisualization System,” Reports of the Machine Learning and Inference Laboratory, MLI95-5, George Mason University, Fairfax, VA, 1995.

Wnek, J., Kaufman, K., Bloedorn, E. and Michalski, R.S., “Selective Induction LearningSystem AQ15c: The Method and User’s Guide,” Reports of the Machine Learning andInference Laboratory, MLI 95-4, Machine Learning and Inference Laboratory, GeorgeMason University, Fairfax, VA, 1995.

Wnek, J. and Michalski, R.S., “Hypothesis-driven Constructive Induction in AQ17-HCI:A Method and Experiments,” Machine Learning, 14, pp. 139-168, 1994.

Wnek, J., Sarma, J., Wahab, A. and Michalski, R.S., “Comparing Learning Paradigmsvia Diagrammatic Visualization: A Case Study in Single Concept Learning using Symbolic,Neural Net and Genetic Algorithm Methods,” Reports of the Machine Learning andInference Laboratory, MLI 90-2, Machine Learning and Inference Laboratory, GeorgeMason University, Fairfax, VA, 1990.

Zadeh, L., “Fuzzy Sets,” Information and Control, 8, pp. 338-353, 1965.

Zagoruiko, N.G., “Recognition Methods and Their Application,” Sovietsky Radio,Moscow (in Russian), 1972.

Zhuravlev, Y.I. and Gurevitch, I.B., “Pattern Recognition and Image Recognition,”Chapter in Zhuravlev, Y.I. (ed.), Pattern Recognition, Classification, Forecasting:Mathematical Techniques and their Application. Issue 2, Nauka, Moscow, pp. 5-72 (inRussian), 1989.

Ziarko, W.P. (ed.), Rough Sets, Fuzzy Sets and Knowledge Discovery, Berlin, Springer-Verlag, 1994.

Zytkow, J.M., “Combining Many Searches in the FAHRENHEIT Discovery System,”Proceedings of the Fourth International Workshop on Machine Learning, Irvine, CA, pp.281-287, 1987.