association rule
DESCRIPTION
Association Rule. Lecturer: Dr. Bo Yuan E-mail: [email protected]. Future Store. Smart Shelf. Smart Scale. Overview. Frequent Itemsets Association Rules Sequential Patterns. A Real Example. Market-Based Problems. Finding associations among items in a transactional database. - PowerPoint PPT PresentationTRANSCRIPT
Overview
Frequent Itemsets
Association Rules
Sequential Patterns
2
Future Store
3
A Real Example
4
Market-Based Problems
Finding associations among items in a transactional database.
Items Bread, Milk, Chocolate, Butter …
Transaction (Basket) A non-empty subset of all items
Cross Selling Selling additional products or services to an existing customer.
Bundle Discount
Shop Layout Design Minimum Distance vs. Maximum Distance
“Baskets” & “Items” Sentences & Words
5
Definitions
A transaction is a set of items: T={ia, ib,…,it}
T is a subset of I where I is the set of all possible items.
The dataset D contains a set of transactions.
An association rule is in the form of
A set of items is referred to as itemset.
An itemset containing k items is called k-itemset.
An itemset can be seen as a conjunction of items.
6
QPIQIPQP and , where
Transactions
7
Transactions Items
1 Bread, Jelly, Peanut, Butter2 Bread, Butter3 Bread, Jelly4 Bread, Milk, Butter5 Chips, Milk6 Bread, Chips7 Bread, Milk8 Chips, Jelly
Searching for rules in the form of: Bread Butter
Support of an Itemset
The support of an item (or itemset) X is the percentage of transactions in which that item (or itemset) occurs.
8
Itemset Support Itemset SupportBread 6/8 Bread, Butter 3/8Butter 3/8 …Chips 2/8 Bread, Butter, Chips 0/8Jelly 3/8 …Milk 3/8 Bread, Butter, Chips, Jelly 0/8Peanut 1/8 …
Bread, Butter, Chips, Jelly, Milk 0/8…
Bread, Butter, Chips, Jelly, Milk, Peanut 0/8
Support & Confidence of Association Rule
The support of an association rule XY is the percentage of transactions that contain X and Y.
The confidence of an association rule XY is the ratio of the number of transactions that contain {X, Y} to the number of transactions that contain X.
It can be represented equally as
Conditional probability: P(Y|X)9
Support & Confidence of Association Rule
Support measures how often the rule occurs in the dataset.
Confidence measures the strength of the rule.
10
Transactions Items1 Bread, Jelly, Peanut, Butter2 Bread, Butter3 Bread, Jelly4 Bread, Milk, Butter5 Chips, Milk6 Bread, Chips7 Bread, Milk8 Chips, Jelly
Bread Milk
Support: 2/8
Confidence: 1/3
Milk Bread
Support: 2/8
Confidence: 2/3
Frequent Itemsets and Strong Rules
Support and Confidence are bounded by thresholds: Minimum support σ Minimum confidence Φ
A frequent (large) itemset is an itemset with support larger than σ.
A strong rule is a rule that is frequent and its confidence is higher than Φ.
Association Rule Problem Given I, D, σ and Φ, to find all strong rules in the form of XY.
The number of all possible association rules is huge. Brute force strategy is infeasible. A smart way is to find frequent itemsets first.
11
The Big Picture
Step 1: Find all frequent itemsets.
Step 2: Use frequent itemsets to generate association rules. For each frequent itemset f
• Create all non-empty subsets of f.
For each non-empty subset s of f• Output s (f-s) if support (f) / support (s) > Φ
12
{a, b, c}
a b c
a c b
b c a
a b c
b a c
c a b
The key is to find frequent itemsets.
Myth No. 1
A rule with high confidence is not necessarily plausible.
For example: |D|=10000
#{DVD}=7500
#{Tape}=6000
#{DVD, Tape}=4000
Thresholds: σ=30%, Φ=50%
Support(Tape DVD)= 4000/10000=40%
Confidence(Tape DVD)=4000/6000=66%
Now we have a strong rule: Tape DVD Seems that Tapes will help promote DVDs.
However, P(DVD)=75% > P(DVD | Tape) !!
Tape buyers are less likely to purchase DVDs.13
Myth No. 2
14
TransactionsBread, MilkBread, BatteryBread, ButterBread, HoneyBread, ChipsYogurt, CokeBread, BatteryCookie, Jelly
%75)(%100)|( BreadPBatteryBreadP
Myth No. 3
15
Association ≠ Causality
P(Y|X) is just the conditional probability.
Itemset Generation
16
Ø
A B C D
AB
ABC
ABCD
AC CDBDBCAD
BCDACDABD
Itemset Calculation
17
)(NMWO 12 dM
The Apriori Method
One of the best known algorithms in Data Mining
Key ideas A subset of a frequent itemset must be frequent.
• {Milk, Bread, Coke} is frequent {Milk, Coke} is frequent
The supersets of any infrequent itemset cannot be frequent.• {Battery} is infrequent {Milk, Battery} is infrequent
18
Candidate Pruning
19
Ø
A B C D
AB
ABC
ABCD
AC CDBDBCAD
BCDACDABD
General Procedure
Generate itemsets of a particular size.
Scan the database once to see which of them are frequent.
Use frequent itemsets to generate candidate itemsets of size=size+1.
Iteratively find frequent itemsets with cardinality from 1 to k.
Avoid generating candidates that are known to be infrequent.
Require multiple scans of the database.
Efficient indexing techniques such as Hash function & Bitmap may help.
20
Apriori Algorithm
21
Lk Ck+1
22
{ | , ,| | 1}kX Y X Y L X Y k
1{ | , , }kX p X L p L p X
2 {{ , },{2 }}1 2 ,3L 1 { , , ,1 2 3 5}4,L
3 {{ , , },{ , , },{ , 21 2 1 33 4 , },{ , , },{22 , , }}51 2 34 5C
3 {{1,2,3}}C
Lk Ck+1
23
{ | , , , [1, 1], }k k i i k kX Y X Y L X Y i k X Y
2 {{1,2},{2,3}}L 3 {}C
2 {{1,2},{1,3},{2,3}}L 3 {{1,2,3}}C
2 {{1,2},{1,3}}L 3 {{1,2,3}}C
2 {{1,3},{2,3}}L 3 {}C
Ordered List
Correctness
24
1 1, k kX X L X C
1 1 1
1 1
1 1 1
{ ,..., , }
{ ,..., , }{ ,..., , }
k k k
k k k
k k k
X X X L
X X X LX X X L
Join
1111 },,,...,{ kkkk CXXXX
Demo
25
Clothing Example
26
Clothing Example
27
Clothing Example
28
𝐶𝑜𝑛𝑓𝑖𝑑𝑒𝑛𝑐𝑒 ( 𝐽𝑒𝑎𝑛𝑠→ h𝑆 𝑜𝑒𝑠 )=𝑆𝑢𝑝𝑝𝑜𝑟𝑡 ({ 𝐽𝑒𝑎𝑛𝑠 , h𝑆 𝑜𝑒𝑠 })𝑆𝑢𝑝𝑝𝑜𝑟𝑡 ( { 𝐽𝑒𝑎𝑛𝑠 })
= 7/2014 /20
=50 % ≥𝑐
Real Examples
29
Effective Recommendation
30
Sequential Pattern
31
TimeCustomer
1 2 43 5 6 7 8
Sequence
A sequence is an ordered list of elements where each element is a collection of one or more items.
t=<t1 t2 … tm> is a subsequence of s=<s1 s2 … sn> if there exist integers 1 j1<j2< … <jmn such that t1sj1, t2sj2,…, tmsjm.
32
s t Y/N<{2, 4} {3, 6, 5} {8}> <{2} {3, 6} {8}> Yes<{2, 4} {3, 6, 5} {8}> <{2} {8}> Yes
<{1, 2} {3, 4}> <{1} {2}> No<{2, 4} {2, 4} {2, 5}> <{2} {4}> Yes
Support of Sequence
33
CID Time ItemsA 1 1, 2, 4A 2 2, 3A 3 5B 1 1, 2B 2 2, 3, 4C 1 1, 2C 2 2, 3, 4C 3 2, 4, 5D 1 2D 2 3, 4D 3 4, 5E 1 1, 3E 2 2, 4, 5
Support<{1, 2}> 60%<{2, 3}> 60%<{2, 4}> 80%
<{3} {5}> 80%<{1} {2}> 80%<{2} {2}> 60%
<{1} {2, 3}> 60%<{2} {2, 3}> 60%
<{1, 2} {2, 3}> 60%
Candidate Space
Given: {Milk} {Bread}
2-itemset: {Bread, Milk}
2-sequence: <{Bread, Milk}> <{Bread} {Milk}>, <{Milk} {Bread}> <{Bread} {Bread}>, <{Milk} {Milk}>
Order matters in sequences but not for itemsets.
For 1000 items: 1000×1000+=1499500
The search space is much larger than before.
How to generate candidates efficiently?
34
Candidate Generation
A sequence s1 is merged with another sequence s2 if and only if the subsequence obtained by dropping the first item in s1 is identical to the subsequence obtained by dropping the last item in s2.
35
3-sequences<{1} {2} {3}><{1} {2, 5}>
<{1} {5} {3}><{2} {3} {4}><{2, 5} {3}>
<{3} {4} {5}><{5} {3, 4}>
Candidate<{1} {2} {3} {4}><{1} {2, 5} {3}><{1} {5} {3, 4}>
<{2} {3} {4} {5}><{2, 5} {3, 4}>
Pruning<{1} {2, 5} {3}>
s1
s2
Reading Materials
Text Book
J. Han and M. Kamber, Data Mining: Concepts and Techniques, Chapter 6, Morgan Kaufmann.
Core Papers
J. Han, J. Pei, Y. Yin and R. Mao (2004) “Mining frequent patterns without candidate generation: A frequent-pattern tree approach”. Data Mining and Knowledge Discovery, Vol. 3, pp. 53-87.
R. Agrawal and R. Srikant (1995) “Mining sequential patterns”. In Proceedings of the Eleventh International Conference on Data Engineering (ICDE), pp. 3-14.
R. Agrawal and R. Srikant (1994) “Fast algorithms for mining association rules in large databases”. In Proceedings of the 20th International Conference on Very Large Data Bases (VLDB) , pp. 487-499.
R. Agrawal, T. Imielinski, and A. Swami (1993) “Mining association rules between sets of items in large databases”. In Proceedings of the ACM SIGMOD International Conference on Management of Data, pp. 207-216.
36
Review
What is an itemset?
What is an association rule?
What is the support of an itemset or an association rule?
What is the confidence of an association rule?
What does (not) an association rule tell us?
What is the key idea of Apriori?
What is the framework of Apriori?
What is a good recommendation?
What is the difference between itemsets and sequences?
37
Next Week’s Class Talk
Volunteers are required for next week’s class talk.
Topic : Frequent Pattern Tree
Hints: Avoids the costly database scans. Avoids the costly generation of candidates. Applies partitioning-based, divide-and-conquer method. Data Mining and Knowledge Discovery, 8, 53-87, 2004
Length: 20 minutes plus question time
38