21
MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis, Michael I. Jordan (ICML 2013) Discussion by: Piyush Rai March 28, 2014 MAD-Bayes (Broderick et al) March 28, 2014

MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

MAD-Bayes: MAP-based Asymptotic Derivationsfrom Bayes

Tamara Broderick, Brian Kulis, Michael I. Jordan

(ICML 2013)

Discussion by: Piyush Rai

March 28, 2014

MAD-Bayes (Broderick et al) March 28, 2014

Page 2: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Introduction

Inference in NPBayes models can be computationally prohibitive

Traditional approaches: MCMC, Variational Bayes (VB)

Standard MCMC and VB can’t cope with big data

Lots of recent efforts on scaling up NPBayes for big data

Online/Stochastic methods: Sequential Monte Carlo, Particle MCMC,Stochastic Variational Inference

Parallel/Distributed versions of MCMC or VB

Point Estimation based methods (e.g., this paper)

MAD-Bayes (Broderick et al) March 28, 2014

Page 3: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Point Estimation for NPBayes

Can be a quick-and-dirty way of finding a “reasonable” solution

Point estimates can often be sensible initializers for MCMC/VB

Some examples of point estimation based methods for NPBayes:

Greedy Search for DP mixture model (Wang & Dunson, JCGS 2011)

Beam Search for DP mixture model (Daume III, AISTATS 2007)

Beam Search for IBP (Rai & Daume III, ICML 2011)

Submodular Optimization for IBP (Reed & Ghahramani, ICML 2013)

Small-variance asymptotics

For DP and HDP mixture models (Kulis & Jordan, ICML 2012)For Dependent DP mixture models (Campbell et al, NIPS 2013)For HMM/infinite-HMM (Roychowdhury et al, NIPS 2013)For DP and IBP (Today’s paper)

MAD-Bayes (Broderick et al) March 28, 2014

Page 4: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Paper Summary

Deterministic, efficient, point estimation for two NPBayes models

CRP/DP based Gaussian mixture modelIBP/Beta-Bernoulli process based linear Gaussian model

Original motivation: EM algorithm for inference in Gaussian mixture model(GMM) behaves like k-means as the mixture variances shrink to zero

MAD-Bayes (Broderick et al) March 28, 2014

Page 5: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Paper Summary

Deterministic, efficient, point estimation for two NPBayes models

CRP/DP based Gaussian mixture modelIBP/Beta-Bernoulli process based linear Gaussian model

Original motivation: EM algorithm for inference in Gaussian mixture model(GMM) behaves like k-means as the mixture variances shrink to zero

This paper: Applies low-variance asymptotics on the MAP objective

instead of on specific inference algorithms (EM, Gibbs sampling)

Key observation: The negative log-likelihood of a GMM approachesk-means objective as the covariances of Gaussians tend to zero

Result: Leads to objectives/algorithms for CRP/IBP that are reminiscent tothat of k-means clustering

MAD-Bayes (Broderick et al) March 28, 2014

Page 6: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Clustering

Consider data x1, . . . , xN , where xn ∈ RD

Assume data can be grouped into K+ clusters

Cluster assignments: z1, . . . , zN

znk = 1 if xn belongs to cluster k (znk′ = 0 ∀k ′ 6= k)

Chinese Restaurant Process: Gives a prior on K+ and z1:N,1:K+

MAD-Bayes (Broderick et al) March 28, 2014

Page 7: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Chinese Restaurant Process (CRP)

Analogy: Each data point is a customer, each cluster is a table

First customer sits on a new table

Customer n sits on

An existing table k with probability ∝ Sn−1,k =∑

n−1m=1 zm,k

A new table with probability ∝ θ > 0

Probability of this clustering (the table assignments z1:N,1:K+)

P(z1:N,1:K+) = θK+−1 Γ(θ + 1)

Γ(θ + N)

K+

k=1

(SN,k − 1)!

The above is known as Exchangeable Partition Probability Function (EPPF)

MAD-Bayes (Broderick et al) March 28, 2014

Page 8: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

CRP Gaussian Mixture Model

Assume data is generated from a mixture of K+ Gaussians

Mixture component k has mean µk and variance σ2ID

Likelihood: P(x |z , µ) =∏K

+

k=1

n:zn,k=1 Nor(xn|µk , σ2ID)

Prior on component means: P(µ1:K+) =∏K

+

k=1 Nor(µk |0, ρ2ID)

Prior on cluster assignments: P(z1:N,1:K+) = θK+−1 Γ(θ+1)

Γ(θ+N)

∏K+

k=1(SN,k − 1)!

Goal: Find a point estimate of z and µ by maximizing the posterior

argmaxK+,z,µ

P(z , µ|x) ∝ argminK+,z,µ

− logP(x , z , µ)

MAD-Bayes (Broderick et al) March 28, 2014

Page 9: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

MAP Objective: Small-Variance Asymptotics

Goal: Solve argminK+,z,µ − logP(x , z , µ)

Idea: Take the MAP objective, set θ = exp(−λ2/(2σ2)) and consider limitσ2 → 0 (note that θ → 0 as σ2 → 0)

Therefore −2σ2 logP(x , z , µ) =∑K

+

k=1

n:zn,k=1 ||xn − µk ||2 + (K+ − 1)λ2 + O(σ2 log(σ2))

MAD-Bayes (Broderick et al) March 28, 2014

Page 10: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

DP-means algorithm

Clustering objective function

argminK+,z,µ

K+

k=1

n:zn,k=1

||xn − µk ||2 + (K+ − 1)λ2

Just like k-means, except every new cluster incurs λ2 penalty

The algorithm:

Iterate until no change

For each data point xn:

Compute distance from cluster centers: dnk = ||xn − µk ||2, k = 1, . . . ,K

if mink dnk > λ2, set K = K + 1, zn = K , and µK = xn

otherwise set zn = argmink dnk

Update the cluster centers

The algorithm converges to a local minimum (just like k-means)

MAD-Bayes (Broderick et al) March 28, 2014

Page 11: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Latent Feature Allocation

Consider data x1, . . . , xN , where xn ∈ RD

Assume data can be described using K+ latent features

Latent feature assignments: z1, . . . , zN , where zn ∈ {0, 1}K+

znk = 1 if latent feature k is present in xn

Indian Buffet Process: Gives a prior on K+ and z1:N,1:K+

MAD-Bayes (Broderick et al) March 28, 2014

Page 12: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Indian Buffet Process (IBP)

Analogy: Each data point is a customer, each latent feature is a dish

First customer selects K+1 ∼ Poisson(γ) dishes

Customer n selects

An existing dish k with probability ∝ Sn−1,k =∑

n−1m=1 zm,k

K+n ∼ Poisson(γ/n) new dishes

Probability of this latent feature allocation (dish assignments z1:N,1:K+)

P(z1:N,1:K+) =γK

+

exp{−∑N

n=1γn}

∏H

h=1 Kh!

K+

k=1

S−1N,k

(

N

SN,k

)−1

The above is known as Exchangeable Feature Probability Function (EFPF)

MAD-Bayes (Broderick et al) March 28, 2014

Page 13: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

IBP Linear Gaussian Model

Assume data is additive combination of K+ latent features µ1, . . . , µK+

xn ∼ Nor(xn|K

+∑

k=1

znkµk , σ2ID)

Notation: X = [x1, . . . , xN ]⊤,Z = [z1, . . . , zN ]

⊤,A = [µ1, . . . , µK+ ]⊤

Likelihood: P(X |Z ,A) = 1

(2πσ2)(ND/2) exp{−tr((X−ZA)⊤(X−ZA))

2σ2 }

Prior on latent feature means P(A) =∏

K+

k=1Nor(µk |0, ρ2ID)

Prior on latent feature allocations P(Z ) =γK+

exp{−∑

Nn=1

γn}

∏H

h=1Kh!

K+

k=1 S−1N,k

(

N

SN,k

)−1

Goal: Find a point estimate of Z and A by maximizing the posterior

argmaxK+,Z ,A

P(Z ,A|X ) ∝ argminK+,Z ,A

− log P(X ,Z ,A)

MAD-Bayes (Broderick et al) March 28, 2014

Page 14: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

MAP Objective: Small-Variance Asymptotics

Goal: Solve argminK+,Z ,A − logP(X ,Z ,A)

Set γ = exp(−λ2/(2σ2)) for some constant λ2 and consider limit σ2 → 0

Therefore −2σ2 log P(X ,Z ,A) =tr((X − ZA)⊤(X − ZA)) + K

+λ2 +O(σ2 exp(−λ2/(2σ2))) + O(σ2 log(σ2))

MAD-Bayes (Broderick et al) March 28, 2014

Page 15: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

BP-means algorithm

Feature allocation objective function

argminK+,Z ,A

tr((X − ZA)⊤(X − ZA)) + K+λ2

The algorithm:

Iterate until no change

For n = 1, . . . ,N

For k = 1, . . . ,K+, choose the optimal value (0 or 1) of znk

Let Z ′ equal Z but with one new latent feature K+ + 1 only for this data point(and set A′ = A but with an additional row Xn − ZnA)

If the triplet (K+ + 1,Z ′,A′) has lower objective than (K+,Z ,A), replace thelatter with the former

Set A← (Z⊤Z )−1

Z⊤X

The algorithm converges to a local minimum (just like k-means)

MAD-Bayes (Broderick et al) March 28, 2014

Page 16: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Extension: Collapsed Objective

Idea: Integrate out the latent feature means A from the likelihood

Collapsed likelihood P(X |Z ) for the latent feature model

The objective function after the small-variance asymptotics becomes

argminK+,Z

tr(X⊤(IN − Z (Z⊤Z )−1

Z⊤)X ) + K

+λ2

Note: a similar objective obtained for the DP clustering case as well

MAD-Bayes (Broderick et al) March 28, 2014

Page 17: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Parametric Objective

Prior P(Z ) on latent feature allocations (fixed K ) is:

Limiting behavior of the MAP objective as σ2 → 0

argminZ ,A

tr[(X − ZA)⊤(X − ZA)]

The K -features algorithm:

Repeat until no changeFor n = 1, . . . ,N

For k = 1, . . . ,K , set znk to minimize ||xn − zn,1:KA||2

Set A = (Z⊤Z )−1

Z⊤X

MAD-Bayes (Broderick et al) March 28, 2014

Page 18: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Experiments

Experiments on the IBP based latent feature model

Datasets considered: Tabletop data and Faces data (both are imagedatasets)

Algorithms considered

Gibbs sampling for the IBP (Gibbs)

BP-means (BP-m)

Collapsed BP-means (Collap)

Stepwise K -features (FeatK)

Note: BP-m, Collap, FeatK were run 1000 times with different initializations

Hyperparameter λ2 was set to a “reasonable” value

MAD-Bayes (Broderick et al) March 28, 2014

Page 19: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Tabletop Data

100 color images, each reduced to a 100 dimensional feature vector via PCAExample images and discovered latent features by BP-means

Timing results and number of inferred latent features

MAD-Bayes (Broderick et al) March 28, 2014

Page 20: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Faces Data

400 images of 200 people (2 images for each person - neutral and smiling),each image reduced to a 100 dimensional feature vector via PCA

Only stepwise K -features and K -means compared

A simple demonstration of why K latent features may be better than K (ormore) clusters

MAD-Bayes (Broderick et al) March 28, 2014

Page 21: MAD-Bayes: MAP-based Asymptotic Derivations from Bayeslcarin/Piyush3.28.2014.pdfMar 28, 2014  · MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick, Brian Kulis,

Conclusions

Point estimation algorithms via small-variance asymptotics on the MAPobjective functions

Resulting objectives are akin to those used in other model-selection methodssuch as AIC

Algorithms similar to k-means, easy to implement (and have potential forsome parallelization)

Initialization can be an issue (tricks from k-means literature can be used)

The algorithm critically depends on λ. It is unclear how to set it.

Thanks!

MAD-Bayes (Broderick et al) March 28, 2014