course on data mining (581550-4)

63
21.11.2001 Data mining: KDD Process 1 Intro/Ass. Rules Intro/Ass. Rules Episodes Episodes Text Mining Text Mining Home Exam Home Exam 24./26.10. 30.10. Clustering Clustering KDD Process KDD Process Appl./Summary Appl./Summary 14.11. 21.11. 7.11. 28.11. Course on Data Mining Course on Data Mining (581550-4) (581550-4)

Upload: tanek-ramsey

Post on 01-Jan-2016

15 views

Category:

Documents


4 download

DESCRIPTION

7.11. 24./26.10. 14.11. Home Exam. 30.10. 21.11. 28.11. Course on Data Mining (581550-4). Intro/Ass. Rules. Clustering. Episodes. KDD Process. Text Mining. Appl./Summary. Course on Data Mining (581550-4). Today 22.11.2001. Today's subject : KDD Process Next week's program : - PowerPoint PPT Presentation

TRANSCRIPT

21.11.2001 Data mining: KDD Process 1

Intro/Ass. RulesIntro/Ass. RulesIntro/Ass. RulesIntro/Ass. Rules

EpisodesEpisodesEpisodesEpisodes

Text MiningText MiningText MiningText Mining

Home ExamHome Exam

24./26.10.

30.10.

ClusteringClusteringClusteringClustering

KDD ProcessKDD ProcessKDD ProcessKDD Process

Appl./SummaryAppl./SummaryAppl./SummaryAppl./Summary

14.11.

21.11.

7.11.

28.11.

Course on Data Mining (581550-4)Course on Data Mining (581550-4)

21.11.2001 Data mining: KDD Process 2

Today 22.11.2001Today 22.11.2001Today 22.11.2001Today 22.11.2001

• Today's subjectToday's subject: :

o KDD ProcessKDD Process

• Next week's programNext week's program: :

o Lecture: Data mining Lecture: Data mining applications, future, summaryapplications, future, summary

o Exercise: KDD ProcessExercise: KDD Process

o Seminar: KDD ProcessSeminar: KDD Process

Course on Data Mining (581550-4)Course on Data Mining (581550-4)

21.11.2001 Data mining: KDD Process 3

KDD processKDD process

• OverviewOverview

• PreprocessingPreprocessing

• Post-processingPost-processing

• SummarySummary

OverviewOverview

21.11.2001 Data mining: KDD Process 4

What is KDD? A process!What is KDD? A process!

• Aim:Aim: the selection and processing selection and processing of dataof data for

o the identification of novel, accurate, and useful patterns, and

o the modeling of real-world phenomena

• Data mining is a major Data mining is a major component of the KDD processcomponent of the KDD process

21.11.2001 Data mining: KDD Process 5

Typical KDD processTypical KDD process

Data miningData miningData miningData miningInput dataInput dataInput dataInput data ResultsResultsResultsResultsPreprocessingPreprocessing PostprocessingPostprocessing

OperationalOperationalDatabaseDatabase

OperationalOperationalDatabaseDatabase

Selection

Selection

Selection

Selection

UtilizationUtilizationUtilizationUtilization

CleanedVerifiedFocused

Eval. ofinteres-tingness

Raw data

Target dataset

Selected usable

patterns

1 32

21.11.2001 Data mining: KDD Process 6

Learning the domainLearning the domainLearning the domainLearning the domain

Data reduction and projectionData reduction and projectionData reduction and projectionData reduction and projection

Creating a target data setCreating a target data setCreating a target data setCreating a target data set

Data cleaning, integrationData cleaning, integrationand transformationand transformation

Data cleaning, integrationData cleaning, integrationand transformationand transformation

Choosing the DM taskChoosing the DM taskChoosing the DM taskChoosing the DM task

Pre-Pre-processingprocessing

Phases of the KDD process (1)Phases of the KDD process (1)

21.11.2001 Data mining: KDD Process 7

Choosing the DM algorithm(s)Choosing the DM algorithm(s)Choosing the DM algorithm(s)Choosing the DM algorithm(s)

Knowledge presentationKnowledge presentationKnowledge presentationKnowledge presentation

Data mining: searchData mining: searchData mining: searchData mining: search

Pattern evaluationPattern evaluationand interpretationand interpretationPattern evaluationPattern evaluationand interpretationand interpretation

Use of discovered knowledgeUse of discovered knowledgeUse of discovered knowledgeUse of discovered knowledge

Post-Post-processingprocessing

Phases of the KDD process (2)Phases of the KDD process (2)

21.11.2001 Data mining: KDD Process 8

Preprocessing - overviewPreprocessing - overview

• Why data preprocessing?Why data preprocessing?

• Data cleaning Data cleaning

• Data integration and Data integration and transformationtransformation

• Data reductionData reduction

PreprocessingPreprocessing

21.11.2001 Data mining: KDD Process 9

Why data preprocessing?Why data preprocessing?

• Aim: Aim: to select the data relevant with respect to the task in hand to be mined

• Data in the real world is dirtyData in the real world is dirty

o incompleteincomplete: lacking attribute values, lacking certain attributes of interest, or containing only aggregate data

o noisynoisy: containing errors or outliers

o inconsistentinconsistent: containing discrepancies in codes or names

• No quality data, no quality mining results!No quality data, no quality mining results!

21.11.2001 Data mining: KDD Process 10

Measures of data qualityMeasures of data quality

o accuracyaccuracy

o completenesscompleteness

o consistencyconsistency

o timelinesstimeliness

o believabilitybelievability

o value addedvalue added

o interpretabilityinterpretability

o accessibilityaccessibility

21.11.2001 Data mining: KDD Process 11

Preprocessing tasks (1)Preprocessing tasks (1)

• Data cleaningData cleaning

o fill in missing values, smooth noisy data, identify or remove outliers, and resolve inconsistencies

• Data integrationData integration

o integration of multiple databases, files, etc.

• Data transformationData transformation

o normalization and aggregation

21.11.2001 Data mining: KDD Process 12

Preprocessing tasks (2)Preprocessing tasks (2)

• Data reductionData reduction (including discretizationdiscretization)

o obtains reduced representation in volume, but produces the same or similar analytical results

o data discretization is part of data reduction, but with particular importance, especially for numerical data

21.11.2001 Data mining: KDD Process 13

Data CleaningData CleaningData CleaningData Cleaning

Data IntegrationData IntegrationData IntegrationData Integration

Data TransformationData TransformationData TransformationData Transformation

Data ReductionData ReductionData ReductionData Reduction

Preprocessing tasks (3)Preprocessing tasks (3)

21.11.2001 Data mining: KDD Process 14

Data cleaning tasksData cleaning tasks

• Fill in missing valuesFill in missing values

• Identify outliers and Identify outliers and smooth out noisy datasmooth out noisy data

• Correct inconsistent Correct inconsistent datadata

21.11.2001 Data mining: KDD Process 15

Missing DataMissing Data

• Data is not always availableData is not always available

• Missing data may be due toMissing data may be due to

o equipment malfunction

o inconsistent with other recorded data, and thus deleted

o data not entered due to misunderstanding

o certain data may not be considered important at the time of entry

o not register history or changes of the data

• Missing data may need to be inferredMissing data may need to be inferred

21.11.2001 Data mining: KDD Process 16

How to Handle Missing Data? (1)How to Handle Missing Data? (1)

• Ignore the tupleIgnore the tuple

o usually done when the class label is missing

o not effective, when the percentage of missing values per attribute varies considerably

• Fill in the missing value manuallyFill in the missing value manually

o tedious + infeasible?

• Use a global constant to fill in the missing valueUse a global constant to fill in the missing value

o e.g., “unknown”, a new class?!

21.11.2001 Data mining: KDD Process 17

How to Handle Missing Data? (2)How to Handle Missing Data? (2)

• Use the attribute mean to fill in the missing valueUse the attribute mean to fill in the missing value

• Use the attribute mean for all samples belonging to Use the attribute mean for all samples belonging to the same class to fill in the missing valuethe same class to fill in the missing value

o smarter solution than using the “general” attribute mean

• Use the most probable value to fill in the missing Use the most probable value to fill in the missing valuevalue

o inference-based tools such as decision tree induction or a Bayesian formalism

o regression

21.11.2001 Data mining: KDD Process 18

Noisy DataNoisy Data

• NoiseNoise: random error or variance in a measured variable

• Incorrect attribute values may due toIncorrect attribute values may due to

o faulty data collection instruments

o data entry problems

o data transmission problems

o technology limitation

o inconsistency in naming convention

21.11.2001 Data mining: KDD Process 19

How to Handle Noisy Data?How to Handle Noisy Data?

• BinningBinning

o smooth a sorted data value by looking at the values around it

• ClusteringClustering

o detect and remove outliers

• Combined computer and human inspectionCombined computer and human inspection

o detect suspicious values and check by human

• RegressionRegression

o smooth by fitting the data into regression functions

21.11.2001 Data mining: KDD Process 20

Binning methods (1)Binning methods (1)

• Equal-depth (frequency) partitioningEqual-depth (frequency) partitioning

o sort data and partition into bins, N intervals, each containing approximately same number of samples

o smooth by bin meansbin means, bin medianbin median, bin bin boundariesboundaries, etc.

o good data scaling

o managing categorical attributes can be tricky

21.11.2001 Data mining: KDD Process 21

Binning methods (2)Binning methods (2)

• Equal-width (distance) partitioningEqual-width (distance) partitioning

o divide the range into N intervals of equal size: uniform grid

o if A and B are the lowest and highest values of the attribute, the width of intervals will be: W = (B-A)/N.

o the most straightforward

o outliers may dominate presentation

o skewed data is not handled well

21.11.2001 Data mining: KDD Process 22

Equal-depth binning - ExampleEqual-depth binning - Example

• Sorted data for priceSorted data for price (in dollars): o 4, 8, 9, 15, 21, 21, 24, 25, 26, 28, 29, 34

• Partition intoPartition into (equal-depth) binsbins:o Bin 1: 4, 8, 9, 15o Bin 2: 21, 21, 24, 25o Bin 3: 26, 28, 29, 34

• Smoothing by bin meansSmoothing by bin means:o Bin 1: 9, 9, 9, 9o Bin 2: 23, 23, 23, 23o Bin 3: 29, 29, 29, 29

• ……by bin boundaries:by bin boundaries:o Bin 1: 4, 4, 4, 15o Bin 2: 21, 21, 25, 25o Bin 3: 26, 26, 26, 34

21.11.2001 Data mining: KDD Process 23

Data Integration (1)Data Integration (1)

• Data integrationData integration

o combines data from multiple sources into a coherent store

• Schema integrationSchema integration

o integrate metadata from different sources

o entity identification problementity identification problem: identify real world entities from multiple data sources, e.g., A.cust-id B.cust-#

21.11.2001 Data mining: KDD Process 24

Data Integration (2)Data Integration (2)

• Detecting and resolving data value Detecting and resolving data value conflictsconflicts

o for the same real world entity, attribute values from different sources are different

o possible reasonspossible reasons: different representations, different scales, e.g., metric vs. British units

21.11.2001 Data mining: KDD Process 25

Handling Redundant DataHandling Redundant Data

• Redundant data occur often, when multiple databases Redundant data occur often, when multiple databases are integratedare integrated

o the same attribute may have different names in different databases

o one attribute may be a “derived” attribute in another table, e.g., annual revenue

• Redundant data may be detected by correlation Redundant data may be detected by correlation analysisanalysis

• Careful integration of data from multiple sources mayCareful integration of data from multiple sources may

o help to reduce/avoid redundancies and inconsistencies

o improve mining speed and quality

21.11.2001 Data mining: KDD Process 26

Data TransformationData Transformation

• SmoothingSmoothing: remove noise from data

• AggregationAggregation: summarization, data cube construction

• GeneralizationGeneralization: concept hierarchy climbing

• NormalizationNormalization: scaled to fall within a small, specified range, e.g.,

o min-max normalization

o normalization by decimal scaling

• Attribute/feature constructionAttribute/feature construction

o new attributes constructed from the given ones

21.11.2001 Data mining: KDD Process 27

Data ReductionData Reduction

• Data reductionData reduction

o obtains a reduced representation of the data set that is much smaller in volume

o produces the same (or almost the same) analytical results as the original data

• Data reduction strategiesData reduction strategies

o dimensionality reduction

o numerosity reduction

o discretization and concept hierarchy generation

21.11.2001 Data mining: KDD Process 28

Dimensionality ReductionDimensionality Reduction

• Feature selectionFeature selection (i.e., attribute subset selection):o select a minimum set of features such that the

probability distribution of different classes given the values for those features is as close as possible to the original distribution given the values of all features

o reduce the number of patterns in the patterns, easier to understand

• Heuristic methodsHeuristic methods (due to exponential # of choices):o step-wise forward selectiono step-wise backward eliminationo combining forward selection and backward elimination

21.11.2001 Data mining: KDD Process 29

Initial attribute set:Initial attribute set: {A1, A2, A3, A4, A5, A6}

> Reduced attribute set: {A1, A4, A6}

Dimensionality Reduction - Dimensionality Reduction - ExampleExample

A4 ?

A1? A6?

Class 1 Class 2 Class 1 Class 2

21.11.2001 Data mining: KDD Process 30

Numerosity ReductionNumerosity Reduction

• Parametric methodsParametric methods

o assume the data fits some model, estimate model parameters, store only the parameters, and discard the data (except possible outliers)

o e.g., regression analysis, log-linear models

• Non-parametric methods Non-parametric methods

o do not assume models

o e.g., histograms, clustering, sampling

21.11.2001 Data mining: KDD Process 31

DiscretizationDiscretization

• Reduce the number of values Reduce the number of values for a given continuous attribute continuous attribute by dividing the range of the attribute into intervalsintervals

• Interval labels Interval labels can then be used to replace actual data values

• Some classification algorithms only accept categorical attributes categorical attributes

21.11.2001 Data mining: KDD Process 32

Concept HierarchiesConcept Hierarchies

• Reduce the data Reduce the data by collecting and replacing low level concepts by higher level concepts higher level concepts

• For example, replace numeric values for the attribute age by more general values young, middle-aged, or senior

21.11.2001 Data mining: KDD Process 33

Discretization and Discretization and concept hierarchy generation concept hierarchy generation

for numeric datafor numeric data

• BinningBinning

• Histogram analysisHistogram analysis

• Clustering analysis Clustering analysis

• Entropy-based discretizationEntropy-based discretization

• Segmentation by natural Segmentation by natural partitioningpartitioning

21.11.2001 Data mining: KDD Process 34

Concept hierarchy generation Concept hierarchy generation for categorical datafor categorical data

• Specification of a partial ordering of partial ordering of attributes attributes explicitly at the schema level by users or expertsusers or experts

• Specification of a portion of a a portion of a hierarchy hierarchy by explicit data grouping data grouping

• Specification of a set of attributes, a set of attributes, but not of their partial orderingnot of their partial ordering

• Specification of only a partial set of only a partial set of attributesattributes

21.11.2001 Data mining: KDD Process 35

Specification of Specification of a set of attributesa set of attributes

• Concept hierarchy can be automatically generated based on the number of distinct values per attribute in the given attribute set. The attribute with the most distinct values is placed at the lowest level of the hierarchy.

country

province_or_ state

city

street

15 distinct values

65 distinct values

3567 distinct values

674 339 distinct values

21.11.2001 Data mining: KDD Process 36

Post-processing - overviewPost-processing - overview

• Why data post-Why data post-processing?processing?

• InterestingnessInterestingness

• VisualizationVisualization

• UtilizationUtilization

Post-processingPost-processing

21.11.2001 Data mining: KDD Process 37

Why data post-processing? (1)Why data post-processing? (1)

• Aim:Aim: to show the results, or more precisely the most interesting findings, of the data mining phase to a user/users in an understandable way

• A possible post-processing methodology:A possible post-processing methodology:

o find all potentially interesting patterns according to some rather loose criteria

o provide flexible methods for iteratively and interactively creating different views of the discovered patterns

• Other more restrictive or focused methodologies Other more restrictive or focused methodologies possible as wellpossible as well

21.11.2001 Data mining: KDD Process 38

Why data post-processing? (2)Why data post-processing? (2)

• A post-processing methodology is useful, ifA post-processing methodology is useful, ifo the desired focus is not known in advance (the

search process cannot be optimizedoptimized to look only for the interesting patterns)

o there is an algorithm that can produce all patterns from a class of potentially interesting patterns (the result is completecomplete)

o the time requirement for discovering all potentially interesting patterns is not considerably longer than, if the discovery was focused to a small subset of potentially interesting patterns

21.11.2001 Data mining: KDD Process 39

Are all the discovered pattern Are all the discovered pattern interesting?interesting?

• A data mining system/query may generate thousands of patterns, thousands of patterns, but are they all interesting? interesting?

Usually NOT!Usually NOT!

• How could we then choose choose the interesting patterns?

=> InterestingnessInterestingness

21.11.2001 Data mining: KDD Process 40

Interestingness criteria (1)Interestingness criteria (1)

• Some possible criteria for interestingnessSome possible criteria for interestingness:

o evidence:evidence: statistical significance of finding?

o redundancy:redundancy: similarity between findings?

o usefulness:usefulness: meeting the user's needs/goals?

o novelty:novelty: already prior knowledge?

o simplicity:simplicity: syntactical complexity?

o generality:generality: how many examples covered?

21.11.2001 Data mining: KDD Process 41

Interestingness criteria(2)Interestingness criteria(2)

• One division of interestingness criteria:One division of interestingness criteria:

o objective measures objective measures that are based on statistics and structures of patterns, e.g., J-measure: statistical significance certainty factor: support or frequency strength: confidence

o subjective measures subjective measures that are based on user’s beliefs in the data, e.g., unexpectedness: “is the found pattern surprising?" actionability: “can I do something with it?"

21.11.2001 Data mining: KDD Process 42

Criticism: Support & ConfidenceCriticism: Support & Confidence

• Example: (Aggarwal & Yu, PODS98)Example: (Aggarwal & Yu, PODS98)

o among 5000 students 3000 play basketball, 3750 eat cereal 2000 both play basket ball and eat cereal

o the rule play basketball play basketball eat cereal [40%, 66.7%] eat cereal [40%, 66.7%] is misleading, because the overall percentage of students eating cereal is 75%, which is higher than 66.7%

o the rule play basketball play basketball not eat cereal [20%, not eat cereal [20%, 33.3%]33.3%] is far more accurate, although with lower support and confidence

21.11.2001 Data mining: KDD Process 43

InterestInterest

• Yet anotherYet another objective measureobjective measure for interestingness is interestinterest that is defined as

• Properties of this measure:Properties of this measure:o takes both P(A) and P(B) in consideration:o P(A^B)=P(B)*P(A), if A and B are independent eventso A and B negatively correlated, if the value is less than 1;

otherwise A and B positively correlated.

)()(

)(

BPAP

BAP

21.11.2001 Data mining: KDD Process 44

J-measureJ-measure• Also J-measure J-measure

is an objective measure objective measure for interestingness• Properties of J-measure:Properties of J-measure:

o again, takes both P(A) and P(B) in considerationo value is always between 0 and 1o can be computed using pre-calculated values

))(1

)(1log()(1(

)(

)(log)()(

Bconf

BAconfBAconf

Bconf

BAconfBAconfAconfmeasureJ

21.11.2001 Data mining: KDD Process 45

Support/Frequency/J-measureSupport/Frequency/J-measure

0

500

1000

1500

2000

2500

3000

0 0,1 0,2 0,3 0,4 0,5 0,6 0,7 0,8 0,9 1

Threshold

Ru

les

Dataset 1

Dataset 2

Dataset 3

Dataset 4

Dataset 5

Dataset 6

21.11.2001 Data mining: KDD Process 46

ConfidenceConfidence

0

500

1000

1500

2000

2500

3000

0 0,1 0,2 0,3 0,4 0,5 0,6 0,7 0,8 0,9 1

Confidence threshold

Ru

les

Dataset 1

Dataset 2

Dataset 3

Dataset 4

Dataset 5

Dataset 6

21.11.2001 Data mining: KDD Process 47

Example – Selection of Example – Selection of Interesting Association RulesInteresting Association Rules

• For reducing the number of association rules that have to be considered, we could, for example, use one of the following selection selection criteria:criteria:o frequency and confidenceo J-measure or interesto maximum rule size (whole rule,

left-hand side, right-hand side)o rule attributes (e.g., templates)

21.11.2001 Data mining: KDD Process 48

Example – Example – Problems with selection of rulesProblems with selection of rules

• A rule can correspond to prior knowledge or A rule can correspond to prior knowledge or expectationsexpectations

o how to encode the background knowledge into the system?

• A rule can refer to uninteresting attributes or A rule can refer to uninteresting attributes or attribute combinationsattribute combinations

o could this be avoided by enhancing the preprocessing phase?

• Rules can be redundantRules can be redundant

o redundancy elimination by rule covers etc.

21.11.2001 Data mining: KDD Process 49

Interpretation and evaluation of Interpretation and evaluation of the results of data miningthe results of data mining

• EvaluationEvaluation

o statistical validation and significance testing

o qualitative review by experts in the field

o pilot surveys to evaluate model accuracy

• InterpretationInterpretation

o tree and rule models can be read directly

o clustering results can be graphed and tabled

o code can be automatically generated by some systems

21.11.2001 Data mining: KDD Process 50

Visualization of Visualization of Discovered Patterns (1)Discovered Patterns (1)

• In some cases, visualization of the visualization of the results results of data mining (rules, clusters, networks…) can be very helpful

• Visualization is actually already already important in the preprocessing important in the preprocessing phase phase in selecting the appropriate data or in looking at the data

• Visualization requires training and training and practicepractice

21.11.2001 Data mining: KDD Process 51

Visualization of Visualization of Discovered Patterns (2)Discovered Patterns (2)

• Different backgrounds/usages may require different Different backgrounds/usages may require different forms of representationforms of representationo e.g., rules, tables, cross-tabulations, or pie/bar chart

• Concept hierarchy is also importantConcept hierarchy is also important o discovered knowledge might be more understandable

when represented at high level of abstraction o interactive drill up/down, pivoting, slicing and dicing

provide different perspective to data• Different kinds of knowledge require different kinds of Different kinds of knowledge require different kinds of

representationrepresentationo association, classification, clustering, etc.

21.11.2001 Data mining: KDD Process 52

VisualizationVisualization

21.11.2001 Data mining: KDD Process 53

21.11.2001 Data mining: KDD Process 54

Increasing potentialto supportbusiness decisions End User

Business Analyst

DataAnalyst

DBA

MakingDecisions

Data PresentationVisualization Techniques

Data MiningInformation Discovery

Data Exploration

OLAP, MDA

Statistical Analysis, Querying and Reporting

Data Warehouses / Data Marts

Data SourcesPaper, Files, Information Providers, Database Systems, OLTP

Utilization of the resultsUtilization of the results

21.11.2001 Data mining: KDD Process 55

SummarySummary

• Data mining:Data mining: semi-automatic discovery of interesting patterns from large data sets

• Knowledge discovery is a Knowledge discovery is a process:process:

o preprocessing

o data mining

o post-processing

o using and utilizing the knowledge

21.11.2001 Data mining: KDD Process 56

SummarySummary

• Preprocessing Preprocessing is important in order to get useful results!

• If a loosely defined mining methodology is used, post- post-processing processing is needed in order to find the interesting results!

• Visualization Visualization is useful in pre- and post-processing!

• One has to be able to utilize utilize the found knowledge!

21.11.2001 Data mining: KDD Process 57

References – KDD ProcessReferences – KDD Process

• P. Adriaans and D. Zantinge. Data Mining. Addison-Wesley: Harlow, England, 1996.• R.J. Brachman, T. Anand. The process of knowledge discovery in databases. Advances in

Knowledge Discovery and Data Mining. AAAI/MIT Press, 1996. • D. P. Ballou and G. K. Tayi. Enhancing data quality in data warehouse environments.

Communications of ACM, 42:73-78, 1999.• M. S. Chen, J. Han, and P. S. Yu. Data mining: An overview from a database perspective. IEEE

Trans. Knowledge and Data Engineering, 8:866-883, 1996. • U. M. Fayyad, G. Piatetsky-Shapiro, P. Smyth, and R. Uthurusamy. Advances in Knowledge

Discovery and Data Mining. AAAI/MIT Press, 1996. • T. Imielinski and H. Mannila. A database perspective on knowledge discovery. Communications of

ACM, 39:58-64, 1996.• Jagadish et al., Special Issue on Data Reduction Techniques. Bulletin of the Technical Committee

on Data Engineering, 20(4), December 1997.• D. Keim, Visual techniques for exploring databases. Tutorial notes in KDD’97, Newport Beach, CA,

USA, 1997.• D. Keim, Visual data mining. Tutorial notes in VLDB’97, Athens, Greece, 1997.• D. Keim, and H.P. Krieger, Visual techniques for mining large databases: a comparison. IEEE

Transactions on Knowledge and Data Engineering, 8(6), 1996.

21.11.2001 Data mining: KDD Process 58

References – KDD ProcessReferences – KDD Process

• W. Kloesgen, Explora: A multipattern and multistrategy discovery assistant. In U.M. Fayyad, et al. (eds.), Advances in Knowledge Discovery and Data Mining, 249-271. AAAI/MIT Press, 1996.

• M. Klemettinen, A knowledge discovery methodology for telecommunication network alarm databases. Ph.D. thesis, University of Helsinki, Report A-1999-1, 1999.

• M. Klemettinen, H. Mannila, P. Ronkainen, H. Toivonen, and A.I. Verkamo. Finding interesting rules from large sets of discovered association rules. CIKM’94, Gaithersburg, Maryland, Nov. 1994.

• G. Piatetsky-Shapiro, U. Fayyad, and P. Smith. From data mining to knowledge discovery: An overview. In U.M. Fayyad, et al. (eds.), Advances in Knowledge Discovery and Data Mining, 1-35. AAAI/MIT Press, 1996.

• G. Piatetsky-Shapiro and W. J. Frawley. Knowledge Discovery in Databases. AAAI/MIT Press, 1991..

• D. Pyle. Data Preparation for Data Mining. Morgan Kaufmann, 1999.

• T. Redman. Data Quality: Management and Technology. Bantam Books, New York, 1992.

• A. Silberschatz and A. Tuzhilin. What makes patterns interesting in knowledge discovery systems. IEEE Trans. on Knowledge and Data Engineering, 8:970-974, Dec. 1996.

• D. Tsur, J. D. Ullman, S. Abitboul, C. Clifton, R. Motwani, and S. Nestorov. Query flocks: A generalization of association-rule mining. SIGMOD'98, Seattle, Washington, June 1998.

21.11.2001 Data mining: KDD Process 59

References – KDD ProcessReferences – KDD Process

• Y. Wand and R. Wang. Anchoring data quality dimensions ontological foundations. Communications of ACM, 39:86-95, 1996.

• R. Wang, V. Storey, and C. Firth. A framework for analysis of data quality research. IEEE Trans. Knowledge and Data Engineering, 7:623-640, 1995.

21.11.2001 Data mining: KDD Process 60

Reminder: Course OrganizationReminder: Course Organization

Course EvaluationCourse EvaluationCourse EvaluationCourse Evaluation

• Passing the course: min 30 pointsPassing the course: min 30 pointso home exam: min 13 points (max 30

points)o exercises/experiments: min 8 points

(max 20 points) at least 3 returned and reported

experimentso group presentation: min 4 points (max

10 points)• Remember also the other requirements:Remember also the other requirements:

o attending the lectures (5/7)o attending the seminars (4/5)o attending the exercises (4/5)

21.11.2001 Data mining: KDD Process 61

Seminar Presentations/Groups 9-10Seminar Presentations/Groups 9-10

Visualization and Visualization and data miningdata mining

Visualization and Visualization and data miningdata mining

D. Keim, H.P., Kriegel, T. D. Keim, H.P., Kriegel, T. Seidl: “Supporting Data Seidl: “Supporting Data Mining of Large Databases Mining of Large Databases by Visual Feedback by Visual Feedback Queries", ICDE’94.Queries", ICDE’94.

21.11.2001 Data mining: KDD Process 62

Seminar Presentations/Groups 9-10Seminar Presentations/Groups 9-10

Interestingness Interestingness Interestingness Interestingness

G. Piatetsky-Shapiro, C.J. G. Piatetsky-Shapiro, C.J. Matheus: “The Matheus: “The Interestingness of Interestingness of Deviations”, KDD’94.Deviations”, KDD’94.

21.11.2001 Data mining: KDD Process 63

Thanks to Thanks to Jiawei Han from Simon Fraser University Jiawei Han from Simon Fraser University

andandMika Klemettinen from Nokia Research CenterMika Klemettinen from Nokia Research Center

for their slides for their slides which greatly helped in preparing this lecture! which greatly helped in preparing this lecture!

Also thanks to Also thanks to Fosca Giannotti and Dino Pedreschi from PisaFosca Giannotti and Dino Pedreschi from Pisa

for their slides.for their slides.

KDD processKDD process