Professional Documents
Culture Documents
3
Multiple-Level Association Rules
• Patterns or association rules may have items or
concepts residing at high, low, or multiple
abstraction levels.
Mining Multiple-Level Association
Rules: Concept hierarchy
• Concept hierarchy for the items: a sequence of
mappings from a set of low-level concepts to a
higher-level, more general concept set.
Mining Multiple-Level Association
Rules: A support-confidence
framework
• Top-down strategy
• Counts are accumulated for the calculation of frequent
itemsets at each concept level
• Starting at concept level 1 and working downward in the
hierarchy
• For each level, any algorithm for discovering frequent
itemsets may be used, such as Apriori or its variations.
Mining Multiple-Level Association Rules: Flexible
support settings
7
Multi-level Association: Flexible Support and
Redundancy filtering
• Flexible min-support thresholds: Some items are more valuable but less
frequent
• Use non-uniform, group-based min-support
• E.g., {diamond, watch, camera}: 0.05%; {bread, milk}: 5%; …
9
Mining Multi-Dimensional Association
• Single-dimensional rules:
buys(X, “milk”) Þ buys(X, “bread”)
10
Advanced Frequent Pattern Mining
11
Mining Quantitative Associations
12
Advanced Frequent Pattern Mining
14
Negative and Rare Patterns
• Rare patterns: Very low support but interesting
• E.g., buying Rolex watches
• Mining: Setting individual-based or special group-based
support threshold for valuable items
15
Defining Negative Correlated Patterns (I)
• Definition 1 (support-based)
• If itemsets X and Y are both frequent but rarely occur together, i.e.,
sup(X U Y) < sup (X) * sup(Y)
Then X and Y are negatively correlated, X U Y is a negatively correlated pattern
• If sup(X U Y) < <sup (X) * sup(Y)
Then X and Y are stongly negatively correlated, X U Y is a strongly negatively correlated
pattern
• Problem: A store sold 100 packages A and 100 packages B, only one transaction containing
both A and B.
• When there are in total 200 transactions, we have
s(A U B) = 0.005, s(A) * s(B) = 0.25, s(A U B) < s(A) * s(B)
• When there are 105 transactions, we have
s(A U B) = 1/105, s(A) * s(B) = 1/103 * 1/103, s(A U B) > s(A) * s(B)
• Where is the problem? —Null transactions, i.e., the support-based definition is not null-
invariant!
16
Defining Negative Correlated Patterns (II)
• Definition 2 (Kulzynski measure-based) If itemsets X
and Y are frequent, but (P(X|Y) + P(Y|X))/2 < є, where є
is a negative pattern threshold, then X and Y are
negatively correlated.
• Ex. For the same needle package problem, when no
matter there are 200 or 105 transactions, if є = 0.01, we
have
(P(A|B) + P(B|A))/2 = (0.01 + 0.01)/2 < є
17
Advanced Frequent Pattern Mining
• Summary
18
Constraints in Data Mining
19
Meta-Rule Guided Mining
• Meta-rule can be in the rule form with partially instantiated predicates
and constants
P1(X, Y) ^ P2(X, W) => buys(X, “iPad”)
• The resulting rule derived can be
age(X, “15-25”) ^ profession(X, “student”) => buys(X, “iPad”)
• In general, it can be in the form of
P1 ^ P2 ^ … ^ Pl => Q1 ^ Q2 ^ … ^ Qr
• Method to find meta-rules
• Find frequent (l+r) predicates (based on min-support threshold)
• Push constraint deeply when possible into the mining process (see
the remaining discussions on constraint-push techniques)
• Use confidence, correlation, and other measures when possible
20
Naïve Algorithm: Apriori + Constraint
Database D itemset sup.
L1 itemset sup.
TID Items C1 {1} 2 {1} 2
100 134 {2} 3 {2} 3
200 235 Scan D {3} 3 {3} 3
300 1235 {4} 1 {5} 3
400 25 {5} 3
C2 itemset sup C2 itemset
L2 itemset sup {1 2} 1 Scan D {1 2}
{1 3} 2 {1 3} 2 {1 3}
{2 3} 2 {1 5} 1 {1 5}
{2 3} 2 {2 3}
{2 5} 3
{2 5} 3 {2 5}
{3 5} 2
{3 5} 2 {3 5}
C3 itemset Scan D L3 itemset sup
{2 3 5} {2 3 5} 2
21
Naïve Algorithm: Apriori + Constraint
Database D itemset sup.
L1 itemset sup.
TID Items C1 {1} 2 {1} 2
100 134 {2} 3 {2} 3
200 235 Scan D {3} 3 {3} 3
300 1235 {4} 1 {5} 3
400 25 {5} 3
C2 itemset sup C2 itemset
L2 itemset sup {1 2} 1 Scan D {1 2}
{1 3} 2 {1 3} 2 {1 3}
{2 3} 2 {1 5} 1 {1 5}
{2 3} 2 {2 3}
{2 5} 3
{2 5} 3 {2 5}
{3 5} 2
{3 5} 2 {3 5}
C3 itemset Scan D L3 itemset sup Constraint:
{2 3 5} {2 3 5} 2 Sum{S.price} < 5
22
How can we use rule constraints
to prune the search space?
• pruning pattern search space
• checks candidate patterns and decides whether a
pattern can be pruned
24
Pattern Space Pruning with Anti-Monotonicity Constraints
• Apply: Anti-monotone rules can be pushed into the mining process. Ex. can
be applied at each iteration of Apriori-style algorithms
25
Pattern Space Pruning with Monotonicity Constraints
26
Pattern Space Pruning with Succinctness
• Succinctness:
• Given A1, the set of items satisfying a succinctness constraint C, then
any set S satisfying C is based on A1 , i.e., S contains a subset belonging
to A1
• Idea: Without looking at the transaction database, whether an itemset
S satisfies constraint C can be determined based on the selection of
items
• min(S.Price) > v is succinct
• Sum(S.Price) >v is not succinct
sum(S) £ v ( a Î S, a ³ 0 ) yes no no
sum(S) ³ v ( a Î S, a ³ 0 ) no yes no
range(S) £ v yes no no
range(S) ³ v no yes no
support(S) £ x no yes no
29
Convertible Constraints: Ordering Data in Transactions
TDB (min_sup=2)
TID Transaction
• Convert tough constraints into anti-
10 a, b, c, d, f
monotone or monotone by properly ordering
20 b, c, d, f, g, h
items
30 a, c, d, e, f
• Examine C: avg(S.profit) ³ 25 40 c, e, f, g
• Order items in value-descending order Item Profit
• <a, f, g, d, b, h, c, e> a 40
b 0
• If an itemset afb violates C
c -20
• So does afbh, afb*
d 10
• It becomes anti-monotone! e -30
f 30
g 20
h -10
31
Strongly Convertible Constraints
32
What Constraints Are Convertible?
33
Constraint-Based Frequent Pattern Mining
• Pattern space pruning constraints
34
Advanced Frequent Pattern Mining
• Summary
35
Mining Long Patterns
36
Challenges of mining long
patterns
• The curse of “downward closure” property of frequent
patterns
• Any sub-pattern of a frequent pattern is frequent
• If {a1, a2, …, a100} is frequent, then {a1}, {a2}, …, {a100}, {a1, a2}, {a1,
a3}, …, {a1, a100}, {a1, a2, a3}, … are all frequent! There are about 2100
such frequent itemsets!
• No matter searching in breadth-first (e.g., Apriori) or depth-
first (e.g., FPgrowth), if we still adopt the “small to large”
step-by-step growing paradigm, we have to examine so
many patterns, which leads to combinatorial explosion!
T1 = 2 3 4 ….. 39 40
Example Dataset T2 = 1 3 4 ….. 39 40
: .
: .
min-support σ= 20 T40=1 2 3 4 …… 39
T41= 41 42 43 ….. 79
T42= 41 42 43 ….. 79
# of Mid sized (length=20) closed Patters: æç 40 ö÷
ç 20 ÷ : .
è ø
: .
# of long pattern (length =39): 1 T60= 41 42 43 … 79
α= {41,42,…,79} of size 39
40
Mining Colossal Patterns: Philosophy
41
Pattern-Fusion Strategy: Core Patterns
• Core Patterns
42
Example: Core Patterns
A data set with 400 transactions, if τ = 0.5
A core pattern: (ab) occurs in (abe) 100 times, and (abcef) 100 times, so
Sup(abe)/Sup(ab)=0.5
(abcef) (100) (ab), (ac), (af), (ae), (bc), (bf), (be) (ce), (fe), (e),
(abc), (abf), (abe), (ace), (acf), (afe), (bcf), (bce),
(bfe), (cfe), (abcf), (abce), (bcfe), (acfe), (abfe), (abcef)
43
Robustness of Core Patterns
• A colossal pattern has far more core
patterns than a small-sized pattern
46
Why Is Pattern-Fusion Efficient?
• A bounded-breadth pattern
tree traversal
47
Advanced Frequent Pattern Mining
• Summary
48
Mining Compressed Patterns by
Pattern Clustering
• Clustering is the automatic process of grouping
similar objects together
• The frequent closed patterns are clustered using a
tightness measure called 𝛿 -cluster.
• A representative pattern is selected for each
cluster, thereby offering a compressed version of
the set of frequent patterns.
Mining Compressed Patterns:
Distance Measure
• Let P1 and P2 be two closed patterns. Their
supporting transaction sets are T(P1) and T(P2),
respectively. The pattern distance of P1 and P2, is
defined as
Example. T(P1) ={t1, t2, t3, t4, t5} and T(P2) = {t1, t2, t3, t4, t6}.
The distance between P1 and P2 is
Mining Compressed Patterns:
Expression of patterns
• Given two patterns A and B, we say if ,
where O(A) is the corresponding itemset of A.
• Assume patterns P1,P2, … ,Pk are in the same
cluster, the representative pattern Pr of the cluster
should be able to express all the other patterns in
the cluster.
Mining Compressed Patterns :
𝛿 –clustering
• A pattern P is 𝛿-covered by another pattern P’ if
and
A detailed approach: Liu, Guimei, Haojun Zhang, and Limsoon Wong. "A flexible approach to finding representative pattern sets." IEEE
Transactions on Knowledge and Data Engineering (TKDE) 26.7 (2013): 1562-1574.
Advanced Frequent Pattern Mining
• Summary
57
How to Understand and Interpret Patterns?
Semantic Information
Non-semantic info.
Definitions indicating
semantics
Examples of Usage
Synonyms
Related Words
Semantic annotation of the pattern “frequent,
pattern”
Automatic annotation of frequent
patterns: Context Modeling
• A context unit is a basic object in a database, D,
that carries semantic information and co-occurs
with at least one frequent pattern, p, in at least one
transaction in D.
• A context unit can be an item, a pattern, or even a
transaction, depending on the specific task and data.
• The context of a pattern, p, is a selected set of
weighted context units (referred to as context
indicators) in the database
Semantic Analysis with Context Models
• Summary
64
Summary
65