from frequent itemsets to semantically meaningful visual patterns

35
LOGO From Frequent Itemsets to Semantically Meaningful Visual Patterns Junsong Yuan, Ying Wu, Ming Yang (KDD 2007) Advisor Dr. Koh Jia-Ling Speaker Tu Yi-Lang Date 2008.05.29

Upload: skylar

Post on 14-Jan-2016

27 views

Category:

Documents


0 download

DESCRIPTION

Advisor : Dr. Koh Jia-Ling Speaker : Tu Yi-Lang Date : 2008.05.29. From Frequent Itemsets to Semantically Meaningful Visual Patterns. Junsong Yuan, Ying Wu, Ming Yang (KDD 2007). Introduction. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: From Frequent Itemsets to Semantically Meaningful Visual Patterns

LOGO

From Frequent Itemsets to Semantically Meaningful

Visual Patterns

Junsong Yuan, Ying Wu, Ming Yang (KDD 2007)

Advisor : Dr. Koh Jia-LingSpeaker : Tu Yi-Lang

Date : 2008.05.29

Page 2: From Frequent Itemsets to Semantically Meaningful Visual Patterns

2

Introduction

Data mining techniques that are successful in transaction and text data may not be simply applied to image data that contain high-dimensional features and have spatial structures.

Above all, unlike transaction and text data that are composed of discrete elements without ambiguity, visual patterns generally exhibit large variabilities in their visual appearances.

Page 3: From Frequent Itemsets to Semantically Meaningful Visual Patterns

3

Introduction

Given a collection of unlabeled images, the objective of image data mining is to discover semantically meaningful spatial patterns that appear repetitively among the images.

In this paper, we aim at an even more challenging problem: given a category of images, we expect to discover some meaningful patterns that have semantic meanings and can well interpret the

category.

Page 4: From Frequent Itemsets to Semantically Meaningful Visual Patterns

4

Notations and basic concepts

Image in the database is described as a set

of visual primitives. denotes the high-dimensional feature and

{xi, yi} denotes the spatial location of vi..

For each visual primitive vi I, its local spatial neighbors.

Page 5: From Frequent Itemsets to Semantically Meaningful Visual Patterns

5

Notations and basic concepts

The image database can generate a collection of such groups, where each group Gi is associated to a visual primitive vi.

By further quantizing all the high dimensional features into M classes through k-means clustering, a codebook Ω can be obtained.

Every prototype Wk in the codebook Ω= {W1, ...,WM} a visual item.

Page 6: From Frequent Itemsets to Semantically Meaningful Visual Patterns

6

Notations and basic concepts

Because each visual primitive is uniquely assigned to one of the visual items Wi, the group Gi can be transferred into a transaction Ti.

.

Page 7: From Frequent Itemsets to Semantically Meaningful Visual Patterns

7

Notations and basic concepts

Given the visual item codebook Ω , a set P Ω is called a visual itemset.Let T(P) denote the set of all the occurrences o

f P in T, and the frequency of P is denoted as: .

For a given threshold θ, called a minimum support, itemset P is frequent if frq(P) > θ.

Page 8: From Frequent Itemsets to Semantically Meaningful Visual Patterns

8

Notations and basic concepts

If an itemset P appears frequently, then all of its sub-sets P’ P will also appear frequently, i.e. frq(P) > θ frq(P’) > θ.

.

Page 9: From Frequent Itemsets to Semantically Meaningful Visual Patterns

9

System overview

Figure 1: Overview for meaningful visual pattern discovery.

Page 10: From Frequent Itemsets to Semantically Meaningful Visual Patterns

10

Visual Primitive Extraction

Here applies the PCA-SIFT points as the visual primitives.

Principal component analysis (PCA) is applied to reduce the dimensionality of the feature, finally each visual primitive is described as a 35-dimensional feature vector .

Page 11: From Frequent Itemsets to Semantically Meaningful Visual Patterns

11

Meaningful Itemset Mining

To evaluate the significance of an itemset P Ω, simply checking its frequency frq(P) in T is f

ar from sufficient.Likelihood ratio test criterion

Based on the two hypotheses :• H0: occurrences of P are randomly generated.• H1: occurrences of P are generated by the hidden

pattern.

.

Page 12: From Frequent Itemsets to Semantically Meaningful Visual Patterns

12

Meaningful Itemset Mining

This result indicates that P2 is a more significant pattern than P1 but counter intuitive, this observation challenges our intuition because P2 is not a cohesive pattern.

Page 13: From Frequent Itemsets to Semantically Meaningful Visual Patterns

13

Meaningful Itemset Mining

For a high-order itemset P (|P| > 2), we perform the Student t-test for each pair of its items to check if items Wi and Wj (Wi,Wj P) are really dependent.

A high-order itemset Pi is meaningful only if all of its pairwise subsets can pass the test individually: i, j P, t({Wi,Wj}) > τ, where τ is the confidence threshold for the t-test.

Page 14: From Frequent Itemsets to Semantically Meaningful Visual Patterns

14

Meaningful Itemset Mining

.

Page 15: From Frequent Itemsets to Semantically Meaningful Visual Patterns

15

Spatial Dependency

Page 16: From Frequent Itemsets to Semantically Meaningful Visual Patterns

16

Meaningful Itemset Mining

.

Page 17: From Frequent Itemsets to Semantically Meaningful Visual Patterns

17

Self-supervised Clustering of Visual Item Codebook

Toward discovering meaningful visual patterns in images, it is critical to obtain optimal visual item codebook Ω.

In the unsupervised clustering setting, there does not exist apparent supervisions.

There are hidden patterns that bring structures in the visual primitive distributions.

Page 18: From Frequent Itemsets to Semantically Meaningful Visual Patterns

18

For example, if we observe that item Wi always appears together with item Wj in a local region, we can infer that they should be generated from a hidden pattern rather than randomly generated.

When such hidden patterns (structures) of the data are discovered through our meaningful itemsets mining, we can apply them as supervision to further improve the clustering results.

Self-supervised Clustering of Visual Item Codebook

Page 19: From Frequent Itemsets to Semantically Meaningful Visual Patterns

19

Self-supervised Clustering of Visual Item Codebook

.

The original Ω can be partitioned into two disjoined subsets : . . : meaningful item codebook. . : meaningless item codebook.

Page 20: From Frequent Itemsets to Semantically Meaningful Visual Patterns

20

Self-supervised Clustering of Visual Item Codebook

Neighborhood Component Analysis (NCA) NCA targets at learning a global linear projecti

on matrix A for the original features. Given two visual primitives vi and vj , NCA lear

ns a new metric A and the distance in the transformed space is :

• .

Page 21: From Frequent Itemsets to Semantically Meaningful Visual Patterns

21

Under the above stochastic selection rule of nearest neighbors, NCA tries to maximize the expected number of points correctly classified under the nearest neighbor classifier.

After obtaining the projection matrix A, we update all the visual features of from to , and re-cluster the visual primitives based on their new features .

Self-supervised Clustering of Visual Item Codebook

Page 22: From Frequent Itemsets to Semantically Meaningful Visual Patterns

22

Pattern Summarization of Meaningful Itemsets

Page 23: From Frequent Itemsets to Semantically Meaningful Visual Patterns

23

Pattern Summarization of Meaningful Itemsets

There are many methods to measure their similarity, we apply the Jaccard distance.

Given a collection of meaningful itemsets Ψ = {Pi}, we want to cluster them into unjoined K-clusters.

Each cluster is defined as a meaningful visual pattern, where and .

Page 24: From Frequent Itemsets to Semantically Meaningful Visual Patterns

24

Pattern Summarization of Meaningful Itemsets

Normalized Cut (NCut) Let G = {V,E} denote a fully connected graph,

where each vertex Pi V is an MI, and the weight s(i, j) on each edge represents similarity between two MIs Pi and Pj .

Normalized cut can partition the graph G into clusters.

Page 25: From Frequent Itemsets to Semantically Meaningful Visual Patterns

25

Pattern Summarization of Meaningful Itemsets

In the case of bipartition, V is partitioned into two disjoined sets A∪B = V.

The following cut value needs to be minimized to get the optimal partition :

• .

• Where and

.

Page 26: From Frequent Itemsets to Semantically Meaningful Visual Patterns

26

Pattern Summarization of Meaningful Itemsets

Page 27: From Frequent Itemsets to Semantically Meaningful Visual Patterns

27

Experiments

All the experiments were conducted on a Pentium-4 3.19GHz PC with 1GB RAM running window XP.

To test whether our MIM algorithm can output meaningful patterns, we want to check if the discovered MI are associated with the frequently appeared foreground objects while not located in the clutter backgrounds.

Page 28: From Frequent Itemsets to Semantically Meaningful Visual Patterns

28

Experiments

The following two criteria are proposed for the evaluation : The precision of Ψ: denotes the percentage

of discovered meaningful itemsets Pi Ψ that are located in the foreground objects.

The precision of : denotes the percentage of meaningless items Wi that are located in the background.

In the ideal situation, if .

Page 29: From Frequent Itemsets to Semantically Meaningful Visual Patterns

29

Experiments

Page 30: From Frequent Itemsets to Semantically Meaningful Visual Patterns

30

Experiments

Using retrieval rate η to denote the percentage of retrieved images that contain at least one MI.

.

Page 31: From Frequent Itemsets to Semantically Meaningful Visual Patterns

31

Experiments

By taking advantage of the FP-growth algorithm for closed FIM, our pattern discovery is very efficient.

It thus provides us a powerful tool to explore large object category database where each image contains hundreds of primitive visual features.

Page 32: From Frequent Itemsets to Semantically Meaningful Visual Patterns

32

Experiments

Page 33: From Frequent Itemsets to Semantically Meaningful Visual Patterns

33

Experiments

Recall = # detects / (# detects + # miss detects).

Precision = # detects /( # detects + # false alarms).

Page 34: From Frequent Itemsets to Semantically Meaningful Visual Patterns

34

Experiments

Examples of meaningful itemsets from car category (6 out of 123 images).

The next examples of meaningful itemsets from face category (12 out of 435 images).

Page 35: From Frequent Itemsets to Semantically Meaningful Visual Patterns

35

Experiments