supplementary materials: weighted substructure mining for

4
Supplementary Materials: Weighted Substructure Mining for Image Analysis Sebastian Nowozin, Koji Tsuda Max Planck Institute for Biological Cybernetics Spemannstrasse 38, 72076 T ¨ ubingen, Germany {sebastian.nowozin,koji.tsuda}@tuebingen.mpg.de Takeaki Uno National Institute of Informatics, 2-1-2 Hitotsubashi, Chiyoda-ku, Tokyo, 101-8430, Japan [email protected] Taku Kudo Google Japan Inc., Cerulean Tower 6F, 26-1 Sakuragaoka-cho, Shibuya-ku, Tokyo, 150-8512, Japan [email protected] okhan Bakır Google GmbH, Freigutstrasse 12, 8002 Zurich, Switzerland [email protected] 1. LPBoost Some details have been omitted from the main presenta- tion for space reasons and clarity of presentation. Here we additionally provide a description of the LPBoost algorithm and a detailed motivation and derivation of the 1.5-class ν - LPBoost variant. We also provide a larger image of the unsupervised rank- ing results. 1.1. LPBoost Algorithm The LPBoost algorithm is summarized in Algorithm 1. We use H to denote the space of possible hypothesis. For weighted substructure mining applications this is H = {h(·; t)|(t) ∈T×Ω}. We denote by h i the hypothesis selected at iteration i. Algorithm 1 Linear Programming Boosting (LPBoost) Input: Training set X = {x 1 ,..., x }, x i ∈X , labels Y = {y 1 ,...,y }, y i ∈ {-1, 1}. convergence threshold θ 0. Output: The classification function f (x): X→ R. 1: λ n 1 , n =1,...,‘ 2: γ 0,J 1 3: loop 4: ˆ h argmax h∈H n=1 y n λ n h(x n ) 5: if n=1 y n λ n ˆ h(x n ) γ + θ then 6: break 7: end if 8: h J ˆ h 9: J J +1 10: (λ) solution to the dual of the LP problem, where γ is the objective function value. 11: α Lagrangian multipliers of solution to dual LP problem 12: end loop 13: f (x) := sign J j=1 α j h j (x) 1.2. 1.5-class LPBoost In Figures 1-3 the behaviour of the 1-class, 2-class and 1.5-class classifiers is shown schematically for a 2D toy ex- ample. Given a set of positive samples X 1 = {x 1,1 ,..., x 1,N }

Upload: others

Post on 07-May-2022

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Supplementary Materials: Weighted Substructure Mining for

Supplementary Materials: Weighted Substructure Mining for Image Analysis

Sebastian Nowozin, Koji TsudaMax Planck Institute for Biological CyberneticsSpemannstrasse 38, 72076 Tubingen, Germany

sebastian.nowozin,[email protected]

Takeaki UnoNational Institute of Informatics, 2-1-2 Hitotsubashi, Chiyoda-ku, Tokyo, 101-8430, Japan

[email protected]

Taku KudoGoogle Japan Inc., Cerulean Tower 6F, 26-1 Sakuragaoka-cho, Shibuya-ku, Tokyo, 150-8512, Japan

[email protected]

Gokhan BakırGoogle GmbH, Freigutstrasse 12, 8002 Zurich, Switzerland

[email protected]

1. LPBoostSome details have been omitted from the main presenta-

tion for space reasons and clarity of presentation. Here weadditionally provide a description of the LPBoost algorithmand a detailed motivation and derivation of the 1.5-class ν-LPBoost variant.

We also provide a larger image of the unsupervised rank-ing results.

1.1. LPBoost Algorithm

The LPBoost algorithm is summarized in Algorithm 1.We use H to denote the space of possible hypothesis. Forweighted substructure mining applications this is H =h(·; t, ω)|(t, ω) ∈ T ×Ω. We denote by hi the hypothesisselected at iteration i.

Algorithm 1 Linear Programming Boosting (LPBoost)Input: Training set X = x1, . . . ,x`,xi ∈ X ,

labels Y = y1, . . . , y`, yi ∈ −1, 1.convergence threshold θ ≥ 0.

Output: The classification function f(x) : X → R.1: λn ← 1

` ,∀n = 1, . . . , `2: γ ← 0, J ← 13: loop4: h← argmaxh∈H

∑`n=1 ynλnh(xn)

5: if∑`n=1 ynλnh(xn) ≤ γ + θ then

6: break7: end if8: hJ ← h9: J ← J + 1

10: (λ, γ) ← solution to the dual of the LP problem,where γ is the objective function value.

11: α ← Lagrangian multipliers of solution to dual LPproblem

12: end loop13: f(x) := sign

(∑Jj=1 αjhj(x)

)

1.2. 1.5-class LPBoost

In Figures 1-3 the behaviour of the 1-class, 2-class and1.5-class classifiers is shown schematically for a 2D toy ex-ample.

Given a set of positive samples X1 = x1,1, . . . ,x1,N

Page 2: Supplementary Materials: Weighted Substructure Mining for

negative

positive

Figure 1. 2-class classifier: Learning a sep-aration of positive and negative samples infeature space.

positive

Figure 2. 1-class classifier: Learning a de-scription of the positive class in featurespace using the positive training samplesonly.

positive

Figure 3. 1.5-class classifier: Learning adescription of the positive class in featurespace using both positive and negative train-ing samples.

+−

0 ρ2 ρ1

Figure 4. Separation as achieved by the 1.5-class LPBoost formu-lation: the positive samples are separated from the origin by ρ1,while the output for negative samples is kept below ρ2.

and a set of negative samples X2 = x2,1, . . . ,x2,M, wederive the following new “1.5-class LPBoost” formulation.

min ρ2 − ρ1 +1νN

N∑n=1

ξ1,n +1νM

M∑m=1

ξ2,m (1)

sb.t.∑t∈T

αth(x1,n; t) ≥ ρ1 − ξ1,n, n = 1, . . . , N∑t∈T

αth(x2,m; t) ≤ ρ2 + ξ2,m, m = 1, . . . ,M∑t∈T

αt = 1,

α ∈ R|T |+ , ρ1, ρ2 ∈ R+, ξ1 ∈ RN+ , ξ2 ∈ RM+

where we directly maximize a soft-margin (ρ1 − ρ2) thatseparates positive from negative training samples as illus-trated in Figure 4. The hypotheses are decision stumps thatreward the presence of a pattern:

h(x; t) =

1 t ⊆ x0 otherwise. (2)

The class decision function is given by thresholding atthe margin’s center (ρ1 + ρ2)/2, such that

f(x) = sign

(∑t∈T

αth(x; t)− ρ1 + ρ2

2

). (3)

Problem (1) can be solved by the LPBoost algorithm [1]

using the following dual LP problem.

maxλ,µ,γ

−γ (4)

sb.t.N∑n=1

λnh(x1,n; t)−M∑m=1

µmh(x2,m; t) ≤ γ,

t ∈ T (5)N∑n=1

λn ≥ 1

M∑m=1

µm ≤ 1

0 ≤ λn ≤1νN

, n = 1, . . . , N

0 ≤ µm ≤1νM

, m = 1, . . . ,M.

For solving Problem (4) we again use column-generationtechniques, incrementally adding the most violated con-straint.

Similarly to the original 2-class LPBoost formulation wederive the gain function from the constraints on the hypothe-ses outputs of the dual of (1) to obtain

h = argmaxh∈H

[N∑n=1

λnh(x1,n)−M∑m=1

µmh(x2,m)

], (6)

which is the same as for the 2-class case, except that theset of samples are explicitly split into two sums. For per-forming weighted substructure mining efficiently we needa bound on the gain for a pattern t′. The bound shall beevaluated using only t, where t ⊆ t′ is subpattern of t′; thisallows efficient pruning in the mining algorithm. For thenew formulation we derive the following new bound on thegain function. Using the anti-monotonicity property [2] for

Page 3: Supplementary Materials: Weighted Substructure Mining for

any t ⊆ t′ we have

gain(t′) =N∑n=1

λnh(x1,n; t′)−M∑m=1

µmh(x2,m; t′)

=N∑n=1

λnI(t′ ⊆ x1,n)−M∑m=1

µmI(t′ ⊆ x2,m)

≤N∑n=1

λnI(t′ ⊆ x1,n)

≤N∑n=1

λnI(t ⊆ x1,n).

A drawback of the new formulation (1) is the violation ofthe closed-under-complementation assumption of Demirizet al. [1], hence we are not guaranteed to obtain the op-timal solution (H,α) among all possible sets and weight-ings. In practice this never caused any problems and theconvergence behavior measured by an independent test er-ror is very similar to the 2-class case.

The 1.5-class formulation (1) is a generalization of 1-class ν-Boosting in Ratsch et al. [3] and we recover theoriginal formulation when M = 0, that is, when no neg-ative samples are available.

References[1] A. Demiriz, K. P. Bennett, and J. Shawe-Taylor. Linear pro-

gramming boosting via column generation. Journal of Ma-chine Learning, 46:225–254, 2002.

[2] S. Morishita. Computing optimal hypotheses efficiently forboosting. In S. Arikawa and A. Shinohara, editors, Progressin Discovery Science, volume 2281 of Lecture Notes in Com-puter Science, pages 471–481. Springer, 2002.

[3] G. Ratsch, B. Scholkopf, S. Mika, and K.-R. Muller. SVMand boosting: One class. Technical report, 2000.

Page 4: Supplementary Materials: Weighted Substructure Mining for

Figu

re5.

Top

seve

nm

osti

nflue

ntia

lpat

tern

sfo

rthe

unsu

perv

ised

rank

ing

task

.(T

his

isan

enla

rged

vers

ion

ofth

efig

ure

inth

epa

per.)