21
Text Summarization via Semantic Representation 吳吳吳 2014/07/16

Text Summarization via Semantic Representation

Embed Size (px)

DESCRIPTION

Text Summarization via Semantic Representation. 吳旻誠 2014/07/16. Gist-content Question. Ask the main idea of the talk. The correct answers describe the closest overall theme of the content, and the distractors refer to only small portions of the content. Gist-content Question. - PowerPoint PPT Presentation

Citation preview

Page 1: Text Summarization via Semantic Representation

Text Summarization via Semantic Representation

吳旻誠

2014/07/16

Page 2: Text Summarization via Semantic Representation

Gist-content Question

• Ask the main idea of the talk.

• The correct answers describe the closest overall theme of the content, and the distractors refer to only small portions of the content.

Page 3: Text Summarization via Semantic Representation

Gist-content Question

• Q. Which of the following is closest to the main idea of this talk?– (A)We've had three explanations for why we might

sleep.– (B)When you're tired, and you lack sleep, you have poor

memory, poor creativity, increased impulsiveness, and overall poor judgment.

– (C)If you have good sleep, it increases your concentration, attention, decision-making, creativity, social skills, health.

– (D)You do not do anything much while you're asleep.

Page 4: Text Summarization via Semantic Representation

Gist-content question generation

• Seem the most important sentence as the main idea of the talk.

• Use LexRank to Measure the importance of sentences.

Page 5: Text Summarization via Semantic Representation

LexRank

• Measure the importance of sentences.

• Graph-based Model(Undirected).

• The nodes represent the sentences.

• The edges are the cosine similarity between nodes.

Page 6: Text Summarization via Semantic Representation

LexRank

Page 7: Text Summarization via Semantic Representation

LexRank

Page 8: Text Summarization via Semantic Representation

Conditions that should be satisfied

• Stochastic matrix.

• Irreducible.

• Aperiodic.

Page 9: Text Summarization via Semantic Representation

LexRank

Page 10: Text Summarization via Semantic Representation

Similarity Between Sentences

• But…what’s the similarity between the following sentences?– I will fully support you.– I'll back you up all the way.

Page 11: Text Summarization via Semantic Representation

Deep Learning

• A set of algorithms in Machine Learning area.

• Learning representations of data.

• Has been applied to fields like computer vision , automatic speech recognition, natural language processing.

Page 12: Text Summarization via Semantic Representation

Reduce the Dimensionality of Data with Neural Networks

Page 13: Text Summarization via Semantic Representation

Word2Vec

• A Google open source tool.

• Compute vector representations of words.

• Provides an efficient implementation of– Continuous Bag-of-Words(CBOW) architecture. – Skip-gram architecture.

Page 14: Text Summarization via Semantic Representation

Continuous Bag-of-Words Model(CBOW)

Page 15: Text Summarization via Semantic Representation

Skip-gram Model

Page 16: Text Summarization via Semantic Representation

Softmax

Page 17: Text Summarization via Semantic Representation

Hierarchical Softmax

• Uses a binary tree representation of the output layer with the W words as its leaves.

• Each word w can be reached by an appropriate path from the root of the tree.

Page 18: Text Summarization via Semantic Representation

Sentence Representations

Now we have the representations of words , but how can we represent sentences by these representations? A recursive Deep Learning model was proposed by Stanford Natural Language Processing Group.

Page 19: Text Summarization via Semantic Representation

Recursive Autoencoder

Page 20: Text Summarization via Semantic Representation

Dynamic Pooling