Upload
martin-patrick
View
214
Download
0
Tags:
Embed Size (px)
Citation preview
Latent Boosting for Action RecognitionZhi Feng Huang et al.
BMVC 2011
2014. 6. 12. Jeany Son
Background – learning with latent variables
• Multiple Instance Learning (MI-SVM, mi-SVM)• (-) Single plain latent variable
• Latent SVM • (+) Structured latent variable• (-) Control parameters / Normalize different features
• MILboost• (+) Not require to normalize different features• (-) Single latent variable / Not structured
• hCRF• Learning parameters and weights for features
Latent Boosting: structured latent variable, not require to normalize different features, feature selection
Boosting
• Combining many weak predictors to produce an ensemble predictor• training examples with high error are weighted higher than those with lower
error• Difficult instances get more attention
• AdaBoost : “shortcoming” are identified by high-weight data points
• Gradient Boosting : “shortcomings” are identified by gradients
Gradient Boost
• Gradient Boosting = Gradient Descent + Boosting
• Analogous to line search in steepest descent • Construct the new base-learners to be maximally correlated with the negative
gradient of the loss function, associated with the whole ensemble.• Arbitrary loss functions can be applied
• Function estimate (parametric)• Change the function optimization problem into the parameter estimation one
Function estimation
given
Steepest descent optimization
• “greedy stage-wise” approach of function incrementing with the base-learners
• The optimal step-size rho should be specified at each iteration
• The optimization rule is defined as:
Gradient Boost
• Solution to the parameter estimates can be difficult to obtain
• Choose new function h to be most correlated with –g(x)
• Classic least-squares minimization problem
: Line search by Newton’s method
K-class Gradient Boost
• Goal : learn a set of scoring function
• by minimizing negative log-loss of the training data
• Probability of an example x being class k :
Weak classifier
• Solve the optimization problem
• Select h to the most parallel with the –g(X) by following minimization problem
• Scoring function is updated as
-1
LatentBoost for Human Action Recognition
• A tracklet is denoted by 5 tuples • : image feature• : position of person in the t-th frame of the tracklet
• : latent variables
l1 l2 l3 l4 l5
x1 x2 x3 x4 x5
Features
• Optical flow features (unary)• Split into 4 scalar fields channels & motion magnitude
• Color histogram features (pairwise)• difference between color histograms in rectangular sub-windows taken from
adjacent frames
Positive optical flow features
(a) Bend (b) Jack (c) Jump (d) pJump (e) run (f) side (g) walk (h) wave1 (i) wave2
Latent Boosting
• Assume that • an example (x,y) is associated with a set of latent variables L={l1, l2, …, lT}• These latent variables are constrained by an undirected graph structure
G=(V,E)
• Scoring function of (x,L) pair for the k-th class
where
l1 l2 l3 l4 l5
x1 x2 x3 x4 x5
Weak learners of the unary & pairwise potential
: gradient of loss function w.r.t. unary potential
: gradient of loss function w.r.t. pairwise potential
Marginal distributions
These can be computed efficiently by using Belief Propagation
l1 l2 l3 l4 l5
x1 x2 x3 x4 x5
y
F
Weizmann dataset (83 videos, 9 events)
Typical tracklets (29x60) from the Weizmann dataset
Jacking
Running
Jumping
Waving
TRECVID dataset (5 cameras, 10 videos, 7 events)
• Typical tracklets (29x60) from the TRECVID dataset
Limitations
• Not guaranteed to find the global optimum in a non-convex problem
• Performance of the final classifier is very sensitive to the initialization
• If the latent structure is not a tree, LatentBoost can perform inference with LBP : slow and not exact than BP
• Summation over all the possible latent variable may cause problems
Summary
• Novel boosting algorithm with latent variables
• Applying to the task of human action recognition
• New way to solve problems with a structure of latent variables