frame the crowd: global visual features labeling boosted with crowdsourcing information

14
Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information Presentation: Michael Riegler, AAU Mathias Lux, AAU Christoph Kofler, TU Delft

Upload: michael-riegler

Post on 27-Jun-2015

125 views

Category:

Science


0 download

DESCRIPTION

Presentation of our submission for the Crowdsourcing task of the MediaEval 2013 Workshop.

TRANSCRIPT

Page 1: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Presentation: Michael Riegler, AAU Mathias Lux, AAU Christoph Kofler, TU Delft

Page 2: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Framing• Similar intentions for taking the

pictures will lead to similar framings of the images

Page 3: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Example 1

Page 4: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Example 2

Page 5: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Idea• Solve the problem with a Global Visual

Features approach based on the framing theory– Always available and for free (beside computation time)

• Workers Reliability for Crowdsourcing Information

• Transfer learning

Page 6: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Visual Classifier

• Modification of LIRE framework• Search based• 12 Global features • Feature selection• Feature combination– late fusion

Page 7: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Workers’ Reliability

• Calculate the reliability of a Worker:#(agrees with majority vote) /#(total votes by worker)

• Used as weight for the votes• Together with self report familiarity

as feature vector

Page 8: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Runs1. Reliability measure for workers2. Visual information with MMSys model3. Visual information with low fidelity worker

votes of Fashion10000 dataset model4. Visual information with new, by the

method of run#1, labeled Fashion10000 dataset

5. Visual information based decision for not clear results of run#1

Page 9: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

MediaEval Results

1 2 3 4 50

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

F1 Label 1 F1 Label 2

Crowdsourcing Visual + MMSys Visual + F10000 low Visual + F10k new labeled Crowd + Visual

Page 10: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

MediaEval Results

1 2 3 4 50

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

F1 Label 1 F1 Label 2

Crowdsourcing Visual + MMSys Visual + F10000 low Visual + F10k new labeled Crowd + Visual

Page 11: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Weighted F1 score (WF1)

• Weighted metric of each F1 score per class

• Can help to interpret the results better

• Can compensate differences between biased classes

Page 12: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Cross Validation Results

1 2 3 4 50

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

F1 Label 1 F1 Label 2Weighted F1 Label 1 Weighted F1 Label 2

Crowdsourcing Visual + MMSys Visual + F10000 low Visual + F10k new labeled Crowd + Visual

Page 13: Frame the Crowd: Global Visual Features Labeling boosted with Crowdsourcing Information

Conclusion• Calculating the workers’ reliability performs

well– Well known that metadata leads to better results

• Transfer learning works well– Crowdsourcing can boost visual classification

• With visual features, even small amount of labeled data leads to good results

• Usefulness of Framing is reflected by the results

• Label 1 very good detectable with global visual features, but label 2 not (concept detection)

• Weighted F1 score can help to understand the results better