33
Unsupervised Modeling of Twitter Conversations Alan Ritter (UW) Colin Cherry (NRC) Bill Dolan (MSR)

Unsupervised Modeling of Twitter Conversations

  • Upload
    seamus

  • View
    32

  • Download
    1

Embed Size (px)

DESCRIPTION

Unsupervised Modeling of Twitter Conversations. Alan Ritter (UW) Colin Cherry (NRC) Bill Dolan (MSR). Twitter. Most of Twitter looks like this: I want a cuban sandwich extra bad!! About 10-20% are replies - PowerPoint PPT Presentation

Citation preview

Page 1: Unsupervised Modeling of Twitter Conversations

Unsupervised Modeling of Twitter Conversations

Alan Ritter (UW)Colin Cherry (NRC)

Bill Dolan (MSR)

Page 2: Unsupervised Modeling of Twitter Conversations

Twitter

• Most of Twitter looks like this:– I want a cuban sandwich extra bad!!

• About 10-20% are replies1. I 'm going to the beach this weekend!

Woo! And I'll be there until Tuesday. Life is good.

2. Enjoy the beach! Hope you have great weather!

3. thank you

Page 3: Unsupervised Modeling of Twitter Conversations

Gathering Conversations

• Twitter Public API• Public Timeline– 20 randomly selected posts per minute– Use to get random sample of twitter users• Query to get all their posts• Follow any that are replies to collect conversations

No need for disentanglement[Elsner & Charniak 2008]

Page 4: Unsupervised Modeling of Twitter Conversations

Conversation Length(number of Tweets)

Length Frequency2 9526033 2147874 975335 434476 263497 144328 102779 6180

10 461411 300412 245013 176414 137715 100016 88817 65218 541… …

Page 5: Unsupervised Modeling of Twitter Conversations

Modeling Latent Structure:Dialogue Acts

1. I 'm going to the beach this weekend! Woo! And I'll be there until Tuesday. Life is good.

2. Enjoy the beach! Hope you have great weather!

3. thank you

Status

Comment

Thanks

Page 6: Unsupervised Modeling of Twitter Conversations

Dialogue Acts:Many Useful Applications

• Conversation Agents [Wilks 2006]• Dialogue Systems [Allen et al. 2007]• Dialogue Summarization [Murray et al. 2006]• Flirtation Detection [Ranganath et al. 2009]• …

Page 7: Unsupervised Modeling of Twitter Conversations

Traditional Approaches

• Gather Corpus of Conversations– focus on speech data• Telephone Conversations - [Jurafsky et. al. 1997]• Meetings - [Dhillon et. al. 2004] [Carletta et. al. 2006]

• Annotation Guidelines• Manual Labeling– Expensive

Page 8: Unsupervised Modeling of Twitter Conversations

Dialogue Acts for Internet Conversations

• Lots of Variety– Email [Cohen et. al. 2004]– Internet Forums [Jeong et. al. 2009]– IRC– Facebook– Twitter…– More on the horizon?

• Tags from speech data not always appropriate– Includes: Backchannel, distruption, floorgrabber– Missing: Meeting request, Status post, etc…

Page 9: Unsupervised Modeling of Twitter Conversations

Our Contributions

1. Unsupervised Tagging of Dialogue Acts– First application to open domain

2. Dialogue Modeling on Twitter– Potential for new language technology

applications– Lots of data available

3. Release a dataset of Twitter conversations– Collected over several months– http://research.microsoft.com/en-us/downloads/

8f8d5323-0732-4ba0-8c6d-a5304967cc3f/default.aspx

Page 10: Unsupervised Modeling of Twitter Conversations

1. I 'm going to the beach this weekend! Woo! And I'll be there until Tuesday. Life is good.

2. Enjoy the beach! Hope you have great weather!

3. thank you

Status

Comment

Thanks

Modeling Latent Structure:Dialogue Acts

Page 11: Unsupervised Modeling of Twitter Conversations

Discourse constraints

1. I 'm going to the beach this weekend! Woo! And I'll be there until Tuesday. Life is good.

2. Enjoy the beach! Hope you have great weather!

3. thank you

Status

Comment

Thanks

Page 12: Unsupervised Modeling of Twitter Conversations

Words indicate dialogue act

1. I 'm going to the beach this weekend! Woo! And I'll be there until Tuesday. Life is good.

2. Enjoy the beach! Hope you have great weather!

3. thank you

Status

Comment

Thanks

Page 13: Unsupervised Modeling of Twitter Conversations

Conversation Specific Topic Words

1. I 'm going to the beach this weekend! Woo! And I'll be there until Tuesday. Life is good.

2. Enjoy the beach! Hope you have great weather!

3. thank you

Status

Comment

Thanks

Page 14: Unsupervised Modeling of Twitter Conversations

Content Modeling [Barzilay & Lee 2004]

• Summarization• Model order of events in

news articles– Very specific topics: e.g.

Earthquakes• Model– Sentence Level HMM• states emit whole sentences

– Learn parameters with EM

Where, when

Rictor scale

Damage

Page 15: Unsupervised Modeling of Twitter Conversations

Adapt CM to Unsupervised DA Tagging

Page 16: Unsupervised Modeling of Twitter Conversations

Problem: Strong Topic Clusters

• We want: • Not:

Status

Comment

Thanks

Technology

Food

Page 17: Unsupervised Modeling of Twitter Conversations

Goal: separate content/dialogue words

• Dialogue act words• Conversation specific words• LDA-style topic model– Each word is generated from 1 of 3 sources:

• General English• Conversation Topic Specific Vocabulary• Dialogue Act specific Vocabulary

Similar to: • [Daume III and Marcu, 2006]• [Haghighi & Vanderwende 2009]

Page 18: Unsupervised Modeling of Twitter Conversations

Conversation Model

Page 19: Unsupervised Modeling of Twitter Conversations

Conversation+Topic Model

Page 20: Unsupervised Modeling of Twitter Conversations

Inference• Collapsed Gibbs sampling– Sample each hidden variable conditioned on assignment

of all others– Integrate out parameters

• But, lots of hyperparameters to set– Act transition multinomial– Act emission multinomial– Doc-specific multinomal– English multinomial– Source distribution multinomial

Slice Sampling Hyperparameters [Neal 2003]

Page 21: Unsupervised Modeling of Twitter Conversations

Probability Estimation

• Problem:– Need to evaluate the probability of a conversation– Integrate out hidden variables– Use as a language model

Chibb-style Estimator [Wallach et. al 2009]

[Murray & Salakhutdinov 2009]

Page 22: Unsupervised Modeling of Twitter Conversations

Qualitative Evaluation

• Trained on 10,000 Twitter conversations of length 3 to 6 tweets

Page 23: Unsupervised Modeling of Twitter Conversations

Conversation+Topic modelDialogue act transitions

Page 24: Unsupervised Modeling of Twitter Conversations

Status

Page 25: Unsupervised Modeling of Twitter Conversations

Question

Page 26: Unsupervised Modeling of Twitter Conversations

Question to Followers

Page 27: Unsupervised Modeling of Twitter Conversations

Reference Broadcast

Page 28: Unsupervised Modeling of Twitter Conversations

Reaction

Page 29: Unsupervised Modeling of Twitter Conversations

Evaluation

• How do we know which works best?

• How well can we predict sentence order?• Generate all permutations of a conversation– Compute probability of each– How similar is highest ranked to original order?– Measure permutation similarity with Kendall Tau• Counts number of swaps needed to get desired order

Page 30: Unsupervised Modeling of Twitter Conversations

Bigram LM Baseline EM Conversation Model0

0.05

0.1

0.15

0.2

0.25

0.036

0.22

Tau

Experiments – Conversation Ordering

Page 31: Unsupervised Modeling of Twitter Conversations

Bigram

LM Base

line

EM Conversa

tion Model

Bayesia

n Conversa

tion+Topic0

0.050.1

0.150.2

0.250.3

0.036

0.220.28

Tau

Experiments – Conversation Ordering

Page 32: Unsupervised Modeling of Twitter Conversations

Bigram

LM Base

line

EM Conversa

tion Model

Bayesia

n Conversa

tion+Topic

Bayesia

n Conversa

tion Model

0

0.1

0.2

0.3

0.036

0.220.28 0.31

Tau

Experiments – Conversation Ordering•Content words help predict sentence order

-Adjacent sentences contain similar content words

Page 33: Unsupervised Modeling of Twitter Conversations

Conclusions• Presented a corpus of Twitter Conversations

– http://research.microsoft.com/en-us/downloads/8f8d5323-0732-4ba0-8c6d-a5304967cc3f/default.aspx

• Conversations on Twitter seem to have some common structure• Strong topic clusters are a problem for open-domain

unsupervised DA tagging– Presented an approach to address this

• Gibbs Sampling/Full Bayesian inference seems to outperform EM on Conversation Ordering