51
Improving Automatic Meeting Understanding by Leveraging Meeting Participant Behavior Satanjeev Banerjee, Thesis Proposal. April 21, 2008 1

Improving Automatic Meeting Understanding by Leveraging Meeting Participant Behavior

  • Upload
    ahanu

  • View
    19

  • Download
    0

Embed Size (px)

DESCRIPTION

Improving Automatic Meeting Understanding by Leveraging Meeting Participant Behavior. Satanjeev Banerjee, Thesis Proposal. April 21, 2008. Using Human Knowledge. Knowledge of human experts is used to build systems that function under uncertainty Often captured through in-lab data labeling - PowerPoint PPT Presentation

Citation preview

Page 1: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

1

Improving Automatic Meeting Understanding by Leveraging

Meeting Participant BehaviorSatanjeev Banerjee, Thesis Proposal. April 21, 2008

Page 2: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

2

Using Human Knowledge Knowledge of human experts is used to build

systems that function under uncertainty Often captured through in-lab data labeling

Another source of knowledge: Users of the system Can provide subjective knowledge System can adapt to the users and their information

needs Reduce data needed in the lab

Technical goal: Improve system performance by automatically extracting knowledge from users

Page 3: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

3

Domain: Meetings Problem:

Large parts of meetings contain unimportant information

Some small parts contain important information How to retrieve the important information?

Impact goal: Help humans get information from meetings

(Romano and Nunamaker, 2001)

What information do people need from

meetings?

Page 4: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

4

Understanding Information Needs Survey of 12 CMU faculty members

How often do you need information from past meetings? On average, 1 missed-meeting, 1.5 attended-meeting a month

What information do you need? Missed-meeting: “What was discussed about topic X?” Attended-meeting: Detail question – “What was the

accuracy?”

How do you get the information? From notes if available – high satisfaction If meeting missed – ask face-to-face

(Banerjee, Rose & Rudnicky, 2005)

Task 1: Detect agenda item being discussed

Task 2: Identify utterances to include in notes

Page 5: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

5

Existing Approaches to Accessing Meeting Information

Meeting recording and browsing (Cutler, et al,02), (Ionescu, et al, 02), (Ehlen, et al, 07), (Waibel, et al, 98).

Automatic meeting understanding Meeting transcription (Stolcke, et al, 2004), (Huggins-Daines, et al,

2007) Meeting topic segmentation (Galley, et al, 2003), (Purver, et al,

2006) Activity recognition through vision (Rybski & Veloso, 2004) Action item detection (Ehlen, et al, 07)

Our goalExtract high quality supervision...from meeting participants (best judges of noteworthy info)...during the meeting (when participants are most available)

Classic supervised

learningUnsupervised learning

Meeting participants used after the meeting

Page 6: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

6

Challenges for Supervision Extraction During the Meeting Giving feedback costs the user time and effort Creates a distraction from the user’s main

task – participating in the meeting

Our high-level approach Develop supervision extraction mechanisms

that help meeting participants do their task Interpret participants’ responses as labeled data

Page 7: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

7

Thesis Statement

Develop approaches to extract high quality supervision from system users, by designing extraction mechanisms that help them do their own task, and interpret their actions as labeled data

Page 8: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

8

Roadmap for the Rest of this Talk Review of past strategies for supervision

extraction Approach:

Passive supervision extraction for agenda item labeling

Active supervision extraction to identify noteworthy utterances

Success criteria, contribution and timeline

Page 9: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

9

Past Strategies for Extracting Supervision from Humans Two types of strategies: Passive and Active

Passive: System does not choose which data points user will label E.g.: Improving ASR from user corrections (Burke, et

al, 06)

Active: System chooses which data points user will label E.g.: Have user label traffic images as risky or not

(Saunier, et al, 04)

Past strategies | Passive approach | Active approach | Summary

Page 10: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

10

Research Issue 1: How to Ask Users for Labels? Categorical labels

Associate desktop documents with task label (Shen, et al, 07) Label image of safe roads for robot navigation (Failes & Olsen,

03)

Item scores/rank Rank report items for inclusion in summary (Garera, et al, 07) Pick best schedule from system-provided choices (Weber, et al,

07)

Feedback on features: Tag movies with new text features (Garden, et al, 05) Identify terms that signify document similarity (Godbole, et al,

04)

Past strategies | Passive approach | Active approach | Summary

Page 11: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

11

Research Issue 2: How to Interpret User Actions as Feedback?Depends on similarity between user and system

behavior

Interpretation simple when behaviors are similar E.g.: Email classification (Cohen 96)

Interpretation may be difficult when user behavior and target system behavior are starkly different E.g.: User corrections of ASR output (Burke, et al, 06)

Past strategies | Passive approach | Active approach | Summary

Page 12: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

12

Research Issue 3: How to Select Data Points for Label Query (Active Strategy)? Typical active learning approach:

Goal: Minimize number of labels sought to reach target error

Approach: Choose data points most likely to improve learner E.g.: Pick data points closest to decision boundary

(Monteleoni, et al, 07) Typical assumption: Human’s task is labeling

System user’s task is usually not same as labeling data

Past strategies | Passive approach | Active approach | Summary

Page 13: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

13

Our Overall Approach to Extracting Data from System Users Goal: Extract high quality subjective labeled data

from system users.

Passive approach: Design the interface to ease interpretation of user actions as feedback Task: Label meeting segments with agenda item

Active approach: Develop label query mechanisms that: Query for labels while helping the user do his task Extract labeled data from user actions Task: Identify noteworthy utterances in meetings

Page 14: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

14

Talk Roadmap Review of past strategies for supervision

extraction Approach:

Passive supervision extraction for agenda item labeling

Active supervision extraction to identify noteworthy utterances

Success criteria, contribution and timeline

Page 15: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

15

Passive Supervision: General Approach Goal: Design the interface to enable

interpretation of user actions as feedback Recipe:

Identify kind of labeled data

needed Target a user task

Find relationship between user task and data neededBuild interface for

user task that captures the relationship

Past strategies | Passive approach | Active approach | Summary

Page 16: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

16

Supervision for Agenda Item Detection

Meeting segments labeled with agenda

item1. Most notes refer to discussions in preceding

segment2. A note and its related segment belong to same

agenda item

1. Time stamp speech and notes2. Enable participants to label notes with agenda item

Labeled data

Automatically detect agenda item being discussedUser task

Relationship

Note taking interface

Note taking during meetings

Past strategies | Passive approach | Active approach | Summary

Page 17: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

17

Speech recognition research statusTopic detection research statusFSGs

Insert Agenda

Shared note taking area

Personal notes – not shared

Past strategies | Passive approach | Active approach | Summary

Page 18: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

18

Getting Segmentation from Notes

Note’s time stamp

Note’s agenda item box

100 Speech recognition research status

200 Speech recognition research status

400 Topic detection research status

600 Topic detection research status

800 Speech recognition research status

950 Speech recognition research status

Speech recognition research status

Speech recognition research status

Topic detection research status

300

700

Past strategies | Passive approach | Active approach | Summary

Page 19: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

19

Evaluate the Segmentation How accurate is the extracted segmentation?

Compare to human annotator Also compare to standard topic segmentation algorithms

Evaluation metric: Pk For every pair of time points k seconds apart, ask:

Are the two points in the same segment or not, in the reference? Are the two points in the same segment or not, in the

hypothesis?

Pk =# time pairs where hypothesis and reference disagreeTotal # of time point pairs in the meeting

Past strategies | Passive approach | Active approach | Summary

Page 20: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

20

SmartNotes Deployment in Real Meetings Has been used in 75 real meetings 16 unique participants overall 4 sequences of meetings

Sequence = 3 or more longitudinal meetings

Past strategies | Passive approach | Active approach | Summary

Sequence Num meetings so far

1 (ongoing) 302 273 (ongoing) 84 (ongoing) 4Remaining 6

Page 21: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

21

Data for Evaluation Data: 10 consecutive related meetings

Reference segmentation: Meetings segmented into agenda items by two different annotators.

Inter-annotator agreement: Pk = 0.062

Avg meeting length 31 minutesAvg # agenda items per meeting

4.1

Avg # participants per meeting

3.75 (2 to 5)

Avg # notes per agenda item

5.9

Avg # notes per meeting 25

Past strategies | Passive approach | Active approach | Summary

Page 22: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

22

Results Baseline: TextTiling (Hearst 97) State of the art: (Purver, et al, 2006)

Unsupervised baseline

Segmentation from Smart-Notes data

Purver, et al, 2006

Inter annota-tor agreement

0.39

0.210.26

0.06

P k

Significant Not significant

Past strategies | Passive approach | Active approach | Summary

Page 23: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

23

Does Agenda Item Labeling Help Retrieve Information Faster? 2 10-minute meetings, manually labeled with

agenda items 5 questions prepared for each meeting

Questions prepared without access to agenda items

16 subjects, not participants of the test meetings

Within subjects user study Experimental manipulation: Access to

segmentation versus no segmentation

Past strategies | Passive approach | Active approach | Summary

Page 24: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

24

Minutes to Complete the Task

With agenda item labels Without agenda item labels

7.5

10Significant

Past strategies | Passive approach | Active approach | Summary

Page 25: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

25

Shown So Far Method of extracting meeting segments

labeled with agenda item from note taking Resulting data produces high quality

segmentation Likely to help participants retrieve information

faster

Next: Learn to label meetings that don’t have notes

Past strategies | Passive approach | Active approach | Summary

Page 26: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

26

Proposed Task: Learn to Label Related Meetings that Don’t Have Notes

Plan: Implement language model based detection similar to (Spitters & Kraaiij, 2001). Train agenda item – specific language models on

automatically extracted labeled meeting segments Perform segmentation similar to (Purver, et al, 06) Label new meeting segments with agenda item

whose LM has the lowest perplexity

Past strategies | Passive approach | Active approach | Summary

Page 27: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

27

Proposed Evaluation Evaluate agenda item labeling of meeting with no

notes 3 real meeting sequences with 10 meetings each For each meeting i in each sequence

Train agenda item labeler on automatically extracted labeled data from previous meetings in same sequence

Compute labeling accuracy against manual labels Show improvement in accuracy from meeting to meeting Baseline: Unsupervised segmentation + text matching

between speech and agenda item label text Evaluate effect on retrieving information

Ask users to answer questions from each meeting With agenda item labeling output by improved labeler, versus With agenda item labeling output by baseline labeler

Past strategies | Passive approach | Active approach | Summary

Page 28: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

28

Talk Roadmap Review of past strategies for supervision

extraction Approach:

Passive supervision extraction for agenda item labeling

Active supervision extraction to identify noteworthy utterances

Success criteria, contribution and timeline

Page 29: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

29

Active Supervision System goal: Select data points, and query user for

labels In active learning, human’s task is to provide the labels But system user’s task may be very different from

labeling data

General approach1. Design query mechanisms such that

Each label query also helps the user do his own task The user’s response to the query can be interpreted as a label

2. Choose data points to query by balancing Estimated benefit of query to user Estimated benefit of label to learner

Past strategies | Passive approach | Active approach | Summary

Page 30: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

30

Task: Noteworthy Utterance Detection Goal: Identify noteworthy utterances –

utterances that participants would include in notes

Labeled data needed: Utterances labeled as either “noteworthy” or “not noteworthy”

Past strategies | Passive approach | Active approach | Summary

Page 31: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

31

Extracting Labeled Data Noteworthy utterance detector

Label query mechanism Notes assistance: Suggest utterances for inclusion in

notes during the meeting Helps participants take notes Interpret participants’ acceptances / rejections as

“noteworthy” / “not noteworthy” labels

Method of choosing utterances for suggestion Benefit to user’s note taking Benefit to learner (detector) from user’s

acceptance/rejection

Past strategies | Passive approach | Active approach | Summary

Completed

Proposed

Proposed

Page 32: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

32

Proposed: Noteworthy Utterance DetectorBinary classification of utterances as

noteworthy or not Support Vector Machine classifier Features:

Lexical: Keywords, tf-idf, named entities, numbers Prosodic: speaking rate, f0 max/min Agenda item being discussed Structural: Speaker identity, utterances since last

accepted suggestion Similar to meeting summarization work of (Zhu

& Penn, 2006)

Past strategies | Passive approach | Active approach | Summary

Page 33: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

33

Extracting Labeled Data Noteworthy utterance detector

Label query mechanism Notes assistance: Suggest utterances for inclusion in

notes during the meeting Helps participants take notes Interpret participants’ acceptances / rejections as

“noteworthy” / “not noteworthy” labels

Method of choosing utterances for suggestion Benefit to user’s note taking Benefit to learner (detector) from user’s

acceptance/rejection

Past strategies | Passive approach | Active approach | Summary

Completed

Proposed

Proposed

Page 34: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

34

Mechanism 1: Direct Suggestion

Fix the problem with emailing

Past strategies | Passive approach | Active approach | Summary

Page 35: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

35

Mechanism 2: “Sushi Boat”

pilot testing has been successfulmost participants took twenty minutesron took much longer to finish tasksthere was no crash

Past strategies | Passive approach | Active approach | Summary

Page 36: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

36

Differences between The Mechanisms Direct suggestion

User can provide accept/reject label Higher cost for the user if suggestion is not

noteworthy

Sushi boat suggestion User only provides accept labels Lower cost for the user

Past strategies | Passive approach | Active approach | Summary

Page 37: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

37

Will Participants Accept Suggestions?

Num Offered

Num offered per min

Num accepted

Num accepted per min

% accepted

Direct suggestion

50 0.6 17 0.2 34.0

Sushi boat 273 1.8 85 0.6 31.0

Wizard of Oz study Wizard listened to audio and suggested text 6 meetings – 2 direct mechanism, 4 sushi boat

mechanism

Past strategies | Passive approach | Active approach | Summary

Page 38: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

38

Percentage of Notes from Sushi BoatMeeting

Num lines of notes

Num lines from Sushi boat

% lines from Sushi boat

1 7 6 86%2 24 20 83%3 32 29 91%4 32 30 94%

Total/Avg

95 85 89%

Past strategies | Passive approach | Active approach | Summary

Page 39: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

39

Extracting Labeled Data Noteworthy utterance detector

Label query mechanism Notes assistance: Suggest utterances for inclusion in

notes during the meeting Helps participants take notes Interpret participants’ acceptances / rejections as

“noteworthy” / “not noteworthy” labels

Method of choosing utterances for suggestion Benefit to user’s note taking Benefit to learner (detector) from user’s

acceptance/rejection

Past strategies | Passive approach | Active approach | Summary

Completed

Proposed

Proposed

Page 40: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

40

Method of Choosing Utterances for Suggestion One idea: Pick utterances that either have high

benefit for detector, or high benefit for the user Most beneficial for detector: Least confident

utterances Most beneficial for user: Noteworthy utterances

with high conf Does not take into account user’s past

acceptance pattern Our approach:

Estimate and track user’s likelihood of acceptance Pick utterances that either have high detector

benefit, or is very likely to be acceptedPast strategies | Passive approach | Active approach |

Summary

Page 41: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

41

Estimating Likelihood of Acceptance Features:

Estimated user benefit of suggested utterance

Benefit(utt) =

where T(utt) = time to type utterance, R(utt) = time to read utterance

# suggestions, acceptances, rejections in this and previous meetings

Amount of speech in preceding window of time Time since last suggestion

Combine features using logistic regression Learn per participant from past acceptances/rejections

Past strategies | Passive approach | Active approach | Summary

T(utt) – R(utt)), if utt is noteworthy according to detector– R(utt)), if utt is not noteworthy according to detector

Page 42: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

42 Past strategies | Passive approach | Active approach | Summary

Overall Algorithm for Choosing Utterances for Direct Suggestion

Estimate benefit of

utterance label to detector

Estimate likelihood of acceptance

Given: An utterance and a participantDecision to make: Suggest utterance to participant?

> threshold

?Suggest utterance to

participant

Yes

Don’t sugge

stNo

Combine

Page 43: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

43

Learning Threshold and Combination Wts Train on WoZ data Split meetings into development and test set For each parameter setting

Select utterances for suggestion to user in development set

Compute acceptance rate by comparing against those accepted by the user in the meeting

Of those shown, use acceptances and rejections to retrain utterance detector

Evaluate utterance detector on test set Pick parameter setting with acceptable

tradeoff between utterance detector error rate and acceptance ratePast strategies | Passive approach | Active approach |

Summary

Page 44: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

44

Proposed Evaluation Evaluate improvement in noteworthy utterance

detection 3 real meeting sequences with 15 meetings each Initial noteworthy detector trained on prior data Retrain over first 10 meetings by suggesting notes Test over next 5 Evaluate: After each test meeting, ask participants to

grade automatically identified noteworthy utterances Baseline: Grade utterances identified by prior-trained

detector Evaluate effect on retrieving information

Ask users to answer questions from test meetings With utterances identified by detector trained on 10 meetings, vs. With utterances identified by prior-trained detector

Past strategies | Passive approach | Active approach | Summary

Page 45: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

45

Talk Roadmap Review of past strategies for supervision

extraction Approach:

Passive supervision extraction for agenda item labeling

Active supervision extraction to identify noteworthy utterances

Success criteria, contribution and timeline

Page 46: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

46

Thesis Success Criteria Show agenda item labeling improves with

labeled data automatically extracted from notes Show participants can retrieve information faster

Show noteworthy utterance detection improves with actively extracted labeled data Show participants retrieve information faster

Past strategies | Passive approach | Active approach | Summary

Page 47: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

47

Expected Technical Contribution Framework to actively acquire data labels

from end users

Learning to identify noteworthy utterances by suggesting notes to meeting participants.

Improving topic labeling of meetings by acquiring labeled data from note taking

Past strategies | Passive approach | Active approach | Summary

Page 48: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

48

Summary: Tasks Completed/Proposed

Agenda item detection through passive supervisionDesign interface to acquire labeled data

Completed

Evaluate interface and labeled data obtained

Completed

Implement agenda item detection algorithm

Proposed

Evaluate agenda item detection algorithm

ProposedImportant utterance detection through active learningImplement notes suggestion interface CompletedImplement SVM classifier ProposedEvaluate the summarization Proposed

Past strategies | Passive approach | Active approach | Summary

Page 49: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

49

Proposal TimelineTime frame

Scheduled task

Apr – Jun 08 Iteratively do the following: Continue running Wizard of Oz studies in real meetings to fine-tune label query mechanisms.Analyze WoZ data to identify features for the automatic summarizer

In parallel, implement baseline meeting summarizerJul – Aug 08 Deploy online summarization and notes suggestion

system in real meetings, and iterate on its development based on feedback

Sep – Oct 08

Upon stabilization, perform summarization user study on test meeting groups

Nov 08 Implement agenda detection algorithmDec 08 Perform agenda detection based user studyJan – Mar 09 Write dissertationApr 09 Defend thesis

Past strategies | Passive approach | Active approach | Summary

Page 50: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

50

Thank you!

Page 51: Improving Automatic Meeting Understanding by Leveraging  Meeting Participant Behavior

51

Acceptances Per Participant

Participant Num sushi boat lines accepted

% of acceptances

% of notes in 5 prior meetings

1 87 82.9% 90%2 4 3.8% 10%3 6 5.7% 0%4 8 7.6% Did not attend

Totals 105