20
Viral Marketing – Learning Influence Probabilities

Viral Marketing – Learning Influence Probabilities

  • Upload
    orli

  • View
    33

  • Download
    0

Embed Size (px)

DESCRIPTION

Viral Marketing – Learning Influence Probabilities . Learning Influence M odels. Where do the numbers come from? . Learning influence models. Where do influence probabilities come from? Real world social networks don’t have probabilities! Can we learn the probabilities from action logs? - PowerPoint PPT Presentation

Citation preview

Page 1: Viral Marketing – Learning Influence Probabilities

Viral Marketing – Learning Influence Probabilities

Page 2: Viral Marketing – Learning Influence Probabilities

2

LEARNING INFLUENCE MODELS

Page 3: Viral Marketing – Learning Influence Probabilities

3

Where do the numbers come from?

Page 4: Viral Marketing – Learning Influence Probabilities

4

Learning influence models

• Where do influence probabilities come from?– Real world social networks don’t have probabilities!– Can we learn the probabilities from action logs?– Sometimes we don’t even know the social network– Can we learn the social network, too?

• Does influence probability change over time?– Yes! How can we take time into account?– Can we predict the time at which user is most likely

to perform an action?

Page 5: Viral Marketing – Learning Influence Probabilities

5

Where do the weights come from? • Influence Maximization – Gen 0: academic

collaboration networks (real) with weights assigned arbitrarily using some models: – Trivalency: weights chosen uniformly at random from

{0.1, 0.01, 0.001}.

0.1 0.001

0.01

0.001

0.01 0.01

Page 6: Viral Marketing – Learning Influence Probabilities

6

Where do the weights come from? • Influence Maximization – Gen 0: academic

collaboration networks (real) with weights assigned arbitrarily using some models: – Weighted Cascade:

1/3 1/3

1/3

1/3

1/3 1/3

Other variants: uniform (constant),WC with parallel edges.

Weight assignment not backed by real data.

edges between and LT: IC: prob. Of success of each attempt.

Page 7: Viral Marketing – Learning Influence Probabilities

7

Inference problems• Given a log

• P1. Social network not given– Infer network and edge weights

• P2. Social network given– Infer edge weights

• P3. Social network and attribution given– Explicit “trackbacks” to parent user

– Simple counting

Page 8: Viral Marketing – Learning Influence Probabilities

8

P1. Social network not given• Observe activation times, assume probability of

a successful activation decays (e.g., exponentially) with time

Actual network Learned network

[Gomez-Rodriguez, Leskovec, & Krause KDD 2010]

Page 9: Viral Marketing – Learning Influence Probabilities

9

P2. Social network givenInput data: (1) social graph and (2) action log of past propagations

u12

u45

u45 follows u12

I liked this movie

I read this article

great movie 09:30

09:00

Action Node Time

a u12 1

a u45 2

a u32 3

a u76 8

b u32 1

b u45 3

b u98 7

Page 10: Viral Marketing – Learning Influence Probabilities

10[Saito et al. KES 2008]

P2. Social network given• D(0), D(1), … D(t) nodes that acted at time t.

• cumulative.

• Find that maximizes likelihood

success

failure

Very expensive (not scalable)

Assumes influence weights remain constant over time

Page 11: Viral Marketing – Learning Influence Probabilities

11

P2. Social N/W given • Action log consists of multiple cascades: and

corresp. Cascades may have different lengths --

• Maximize Log Likelihood using EM.

Page 12: Viral Marketing – Learning Influence Probabilities

12

P2. Social network given

• Several models of influence probability– in the context of General Threshold model + time– consistent with IC and LT models

• Models able to predict whether a user will perform an action or not: predict the time at which she will perform it

• Introduce metrics of user and action influenceability – high values genuine influence

• Develop efficient algorithms to learn the parameters of the models; minimize the number of scans over the propagation log

• Incrementality property[[Goyal, Bonchi, and L. WSDM2010 ]

Page 13: Viral Marketing – Learning Influence Probabilities

Generalized Threshold Model • Each node chooses a threshold at random from just like in

LT model.

• But now, the net inflow of influence can be a general function where is the set of in-neighbors of that are currently active.

• activates iff

• Note: Diff. nodes can not only choose diff. they can also use diff. activation functions

• Special Case: LT model.

Page 14: Viral Marketing – Learning Influence Probabilities

Incrementality • In a learning algorithm, need to compute

activation probs of nodes repeatedly as new in-neighbors get activated.

• Suppose we have already computed for a node w.r.t. its current set of active in-neighbors Suppose a new in-neighbor has activated and we want to compute fast, i.e., from and instead of from scratch.

Page 15: Viral Marketing – Learning Influence Probabilities

Incrementality /2 • One activation function that satisifes our

criterion is:

• It’s incremental:

• Furthermore, this function is monotone and submodular. These properties are required since they are assumed by IC/LT models.

• Incrementality is desirable (for speed), not required.

Page 16: Viral Marketing – Learning Influence Probabilities

16

Influence modelsStatic Models: probabilities are static and do not change over time.

Bernoulli: Jaccard:

Continuous Time (CT) Models: probabilities decay exponentially in time

Not incremental, hence very expensive to apply on large datasets.

Discrete Time (CT) Models: Active neighbor u of v remains contagious in

[t, t+ (u,v)], has constant influence prob p(u,v) in the interval and 0 outside.

Monotone, submodular, and incremental!

Page 17: Viral Marketing – Learning Influence Probabilities

17

Evaluation• Flickr groups dataset (action=joining)– ~1.3M nodes, 40M edges, 36M actions– 80/20 training/testing split

• Predict whether user will become active or not, given active neighbors

Reality

Prediction

Active InactiveActive TP FP

Inactive FN TNTotal P N

Operating Point

Ideal Point

Binary Classification(ROC curve)

TPR = TP/P = TP/(TP+FN)

FPR = FP/N = FP/(FP+TN)

Page 18: Viral Marketing – Learning Influence Probabilities

18

Comparison of Static, CT and DT models

• Time-conscious models better than the static model– CT and DT models perform equally well

• Static and DT models are far more efficient compared to CT models because of their incremental nature

Page 19: Viral Marketing – Learning Influence Probabilities

19

Predicting Time – Distribution of Error

• Operating Point is chosen corresponding to – TPR: 82.5%, FPR: 17.5%.

• Most of the time, error in the prediction is very small

Page 20: Viral Marketing – Learning Influence Probabilities

20

Learning Influence Probabilities Takeaways

• Influence network and weights not always available• Learn from the action log

– [Gomez-Rodriguez et al. 2010]: Infer social network and edge weights– [Saito et al. 2008]: Infer edge weights using EM approach– [Goyal et al. 2010]: Infer both static and time-conscious models of

influence• Using CT models, it is possible to predict even the time at

which a user will perform it with a good accuracy.• Introduce metrics of users and actions influenceability.

– High values => easier prediction of influence.– Can be utilized in Viral Marketing decisions.