39
Agenda Agenda • Introduction Bag-of-words models Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based image retrieval Datasets & Conclusions

Agenda

  • Upload
    homer

  • View
    35

  • Download
    0

Embed Size (px)

DESCRIPTION

Agenda. Introduction Bag-of-words models Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based image retrieval Datasets & Conclusions. Classifier based methods. Decision boundary. Background. Computer screen. - PowerPoint PPT Presentation

Citation preview

Page 1: Agenda

AgendaAgenda• Introduction

• Bag-of-words models

• Visual words with spatial location

• Part-based models

• Discriminative methods

• Segmentation and recognition

• Recognition-based image retrieval

• Datasets & Conclusions

Page 2: Agenda

Classifier based methodsObject detection and recognition is formulated as a classification problem.

Bag of image patches

… and a decision is taken at each window about if it contains a target object or not.

Decision boundary

Computer screen

Background

In some feature space

Where are the screens?

The image is partitioned into a set of overlapping windows

Page 3: Agenda

Discriminative methods

106 examples

Nearest neighbor Neural networks

Support Vector Machines and Kernels Conditional Random Fields

Page 4: Agenda

Nearest Neighbors

106 examples

Shakhnarovich, Viola, Darrell 2003

Difficult due to high intrinsic dimensionality of images- lots of data needed- slow neighbor lookup

Torralba, Fergus, Freeman 2008

Page 5: Agenda

Multi-layer Hubel-Wiesel architectures

Neural networks

LeCun, Bottou, Bengio, Haffner 1998Rowley, Baluja, Kanade 1998Hinton & Salakhutdinov 2006Ranzato, Huang, Boureau, LeCun 2007

Riesenhuber & Poggio 1999Serre, Wolf, Poggio. 2005Mutch & Lowe 2006

Biologically inspired

Page 6: Agenda

Support Vector Machines

Heisele, Serre, Poggio, 2001

Face detection

Pyramid Match Kernel

Combining Multiple Kernels

Varma & Roy 2007Bosch, Munoz, Zisserman 2007

Grauman & Darrell 2005Lazebnik, Schmid, Ponce 2006

Page 7: Agenda

Conditional Random FieldsKumar & Hebert 2003

Quattoni, Collins, Darrell 2004

More in Segmentation section

Page 8: Agenda

• A simple algorithm for learning robust classifiers– Freund & Shapire, 1995– Friedman, Hastie, Tibshhirani, 1998

• Provides efficient algorithm for sparse visual feature selection– Tieu & Viola, 2000– Viola & Jones, 2003

• Easy to implement, not requires external optimization tools.

Boosting

Page 9: Agenda

A simple object detector with Boosting Download

• Toolbox for manipulating dataset

• Code and dataset

Matlab code

• Gentle boosting

• Object detector using a part based model

Dataset with cars and computer monitors

http://people.csail.mit.edu/torralba/iccv2005/

Page 10: Agenda

Boosting

Boosting fits the additive model

by minimizing the exponential loss

Training samples

The exponential loss is a differentiable upper bound to the misclassification error.

Page 11: Agenda

Weak classifiers • The input is a set of weighted training

samples (x,y,w)

• Regression stumps: simple but commonly used in object detection.

Four parameters:

b=Ew(y [x> ])

a=Ew(y [x< ])x

fm(x)

Page 12: Agenda

From images to features:A myriad of weak detectors

We will now define a family of visual features that can be used as weak classifiers (“weak detectors”)

Takes image as input and the output is binary response.The output is a weak detector.

Page 13: Agenda

A myriad of weak detectors

• Yuille, Snow, Nitzbert, 1998• Amit, Geman 1998• Papageorgiou, Poggio, 2000• Heisele, Serre, Poggio, 2001• Agarwal, Awan, Roth, 2004• Schneiderman, Kanade 2004 • Carmichael, Hebert 2004• …

Page 14: Agenda

Weak detectors

Textures of textures Tieu and Viola, CVPR 2000

Every combination of three filters generates a different feature

This gives thousands of features. Boosting selects a sparse subset, so computations on test time are very efficient. Boosting also avoids overfitting to some extend.

Page 15: Agenda

Haar wavelets

Haar filters and integral imageViola and Jones, ICCV 2001

The average intensity in the block is computed with four sums independently of the block size.

Page 16: Agenda

Haar waveletsPapageorgiou & Poggio (2000)

Polynomial SVM

Page 17: Agenda

Edges and chamfer distance

Gavrila, Philomin, ICCV 1999

Page 18: Agenda

Edge fragments

Weak detector = k edge fragments and threshold. Chamfer distance uses 8 orientation planes

Opelt, Pinz, Zisserman, ECCV 2006

Page 19: Agenda

Histograms of oriented gradients

• Dalal & Trigs, 2006

• Shape contextBelongie, Malik, Puzicha, NIPS 2000• SIFT, D. Lowe, ICCV 1999

Page 20: Agenda

Weak detectors

Part based: similar to part-based generative models. We create weak detectors by using parts and voting for the object center location

Car model Screen model

These features are used for the detector on the course web site.

Page 21: Agenda

Weak detectors

First we collect a set of part templates from a set of training objects.Vidal-Naquet, Ullman, Nature Neuroscience 2003

Page 22: Agenda

Weak detectors

We now define a family of “weak detectors” as:

= =

Better than chance

*

Page 23: Agenda

Weak detectors

We can do a better job using filtered images

Still a weak detectorbut better than before

* * ===

Page 24: Agenda

Example: screen detectionFeature output

Page 25: Agenda

Example: screen detectionFeature output

Thresholded output

Weak ‘detector’Produces many false alarms.

Page 26: Agenda

Example: screen detectionFeature output

Thresholded output

Strong classifier at iteration 1

Page 27: Agenda

Example: screen detectionFeature output

Thresholded output

Strongclassifier

Second weak ‘detector’Produces a different set of false alarms.

Page 28: Agenda

Example: screen detection

+

Feature output

Thresholded output

Strongclassifier

Strong classifier at iteration 2

Page 29: Agenda

Example: screen detection

+

Feature output

Thresholded output

Strongclassifier

Strong classifier at iteration 10

Page 30: Agenda

Example: screen detection

+

Feature output

Thresholded output

Strongclassifier

Adding features

Finalclassification

Strong classifier at iteration 200

Page 31: Agenda

We want the complexity of the 3 features classifier with the performance of the 100 features classifier:

Cascade of classifiersFleuret and Geman 2001, Viola and Jones 2001

Recall

Precision

0% 100%

100%

3 features

30 features

100 features

Select a threshold with high recall for each stage.

We increase precision using the cascade

Page 32: Agenda

Some goals for object recognition

• Able to detect and recognize many object classes

• Computationally efficient• Able to deal with data starving situations:

– Some training samples might be harder to collect than others

– We want on-line learning to be fast

Page 33: Agenda

Shared features• Is learning the object class 1000 easier

than learning the first?

• Can we transfer knowledge from one object to another?

• Are the shared properties interesting by themselves?

Page 34: Agenda

Shared features

Screen detector

Car detector

Face detector

• Independent binary classifiers:

Torralba, Murphy, Freeman. CVPR 2004. PAMI 2007

Screen detector

Car detector

Face detector

• Binary classifiers that share features:

Page 35: Agenda

50 training samples/class29 object classes2000 entries in the dictionary

Results averaged on 20 runsError bars = 80% interval

Krempp, Geman, & Amit, 2002

Torralba, Murphy, Freeman. CVPR 2004

Shared features

Class-specific features

Page 36: Agenda

Generalization as a function of object similarities

12 viewpoints12 unrelated object classes

Number of training samples per class Number of training samples per class

Are

a un

der R

OC

Are

a un

der R

OC K = 2.1 K = 4.8

Torralba, Murphy, Freeman. CVPR 2004. PAMI 2007

Page 37: Agenda

Sharing patches• Bart and Ullman, 2004

For a new class, use only features similar to features that where good for other classes:

Proposed Dog features

Page 38: Agenda

Sharing transformationsMiller, E., Matsakis, N., and Viola, P. (2000). Learning from one example

through shared densities on transforms. In IEEE Computer Vision and Pattern Recognition.

Transformations are sharedand can be learnt from other tasks.

Page 39: Agenda

Some references on multiclass

• Caruana 1997• Schapire, Singer, 2000• Thrun, Pratt 1997• Krempp, Geman, Amit, 2002• E.L.Miller, Matsakis, Viola, 2000• Mahamud, Hebert, Lafferty, 2001• Fink 2004• LeCun, Huang, Bottou, 2004• Holub, Welling, Perona, 2005• …