i256 applied natural language processing fall 2009

58
1 I256 Applied Natural Language Processing Fall 2009 Text Summarization Barbara Rosario

Upload: sierra

Post on 19-Mar-2016

50 views

Category:

Documents


0 download

DESCRIPTION

I256 Applied Natural Language Processing Fall 2009. Text Summarization. Barbara Rosario. Outline. Introduction and Applications Types of summarization tasks Approaches and paradigms Evaluation methods Acknowledgments Slides inspired (and taken) from - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: I256  Applied Natural Language Processing Fall 2009

1

I256 Applied Natural Language

ProcessingFall 2009

Text Summarization

Barbara Rosario

Page 2: I256  Applied Natural Language Processing Fall 2009

2

Outline• Introduction and Applications• Types of summarization tasks• Approaches and paradigms • Evaluation methods

• Acknowledgments– Slides inspired (and taken) from

• Automatic Summarization by Inderjeet Mani• www.isi.edu/~marcu/acl-tutorial.ppt (Hovy and Marcu)• http://summarization.com/• http://www.summarization.com/sigirtutorial2004.ppt (Radez)

Page 3: I256  Applied Natural Language Processing Fall 2009

3

Introduction• The problem – Information overload

• 4 Billion URLs indexed by Google• 200 TB of data on the Web [Lyman and Varian 03]• Information is created every day in enormous amounts

• One solution – summarization• The goal of an automatic summarization is to

take an information source, extract content from it and present the most important content to the user in a condensed form and in a manner sensitive to the user’s or application’s needs– Other solutions are QA, information extraction, IR,

indexing, document clustering, visualization

Page 4: I256  Applied Natural Language Processing Fall 2009

4

Applications• Abstracts for scientific and other articles• News summarization (mostly multiple document

summarization)– Multimedia news summaries (“watch the news and tell me what

happened while I was away”)• Web pages for search engines• Hand-held devices• Question answering and data/intelligence gathering• Physicians’ aid

– Provide physicians with summaries of on-line medical literature related to a patient’s medical record

• Meetings summarization• Aid for the handicapped

– Summarization for reading machine for the blind.

Page 5: I256  Applied Natural Language Processing Fall 2009

5

• General purpose commercial summarization tools:– AutoSummarize MS Word– InXight Summarizer

Current applications

Page 6: I256  Applied Natural Language Processing Fall 2009

6

MSWord AutoSummarize

Page 7: I256  Applied Natural Language Processing Fall 2009

7

Human summarization and abstracting

• What professional abstractors do– “To take an original article, understand it and

pack it neatly into a nutshell without loss of substance or clarity presents a challenge which many have felt worth taking up for the joys of achievement alone. These are the characteristics of an art form”. (Ashworth)

Page 8: I256  Applied Natural Language Processing Fall 2009

8

Human summarization and abstracting

• The abstract and its use:– To promote current awareness– To save reading time– To facilitate selection– To facilitate literature searches– To improve indexing efficiency– To aid in the preparation of reviews

Page 9: I256  Applied Natural Language Processing Fall 2009

9

American National Standard for Writing Abstracts

– State the purpose, methods, results, and conclusions presented in the original document, either in that order or with an initial emphasis on results and conclusions.

– Make the abstract as informative as the nature of the document will permit, so that readers may decide, quickly and accurately, whether they need to read the entire document.

– Avoid including background information or citing the work of others in the abstract, unless the study is a replication or evaluation of their work.

Cremmins 82, 96

Page 10: I256  Applied Natural Language Processing Fall 2009

10

American National Standard for Writing Abstracts

– Do not include information in the abstract that is not contained in the textual material being abstracted.

– Verify that all quantitative and qualitative information used in the abstract agrees with the information contained in the full text of the document.

– Use standard English and precise technical terms, and follow conventional grammar and punctuation rules.

– Give expanded versions of lesser known abbreviations and acronyms, and verbalize symbols that may be unfamiliar to readers of the abstract.

– Omit needless words, phrases, and sentences.

Cremmins 82, 96

Page 11: I256  Applied Natural Language Processing Fall 2009

11

Types of Summaries • Indicative vs. Informative vs. Critical

– Indicative: give an idea of what is there, provides a reference function for selecting documents for more in-depth reading

– Informative: a substitute for the entire document, covers all the salient information in the source at some level of detail

– Critical: evaluates the subject matter of the source, expressing the abstractor’s view on the quality of the work of the author

Page 12: I256  Applied Natural Language Processing Fall 2009

12

Types of Summaries (cont.) • Input: single document vs. multi-document (MDS)

– MDS: what’s common across documents or different in a particular one

• Input: media types (text, audio, table, pictures, diagrams)

• Output: media types (text, audio, table, pictures, diagrams)

Page 13: I256  Applied Natural Language Processing Fall 2009

13

Types of Summaries (cont.) • Output: Extract vs. Abstract

– Extract: summary consisting entirely of material copied from the input

– Abstract: summary where some material is not present in the input• Paraphrase, generation

– Research shows that sometimes readers prefer extracts

Page 14: I256  Applied Natural Language Processing Fall 2009

14

Types of Summaries (cont.)

• Output: User-focused (or topic-focused or query focused): summaries that are tailored to the requirements of a particular user or group of users– Background

• Does the reader have the needed prior knowledge? Expert reader vs. Novice reader

• General: summaries aimed at a particular –usually broad –readership community

Page 15: I256  Applied Natural Language Processing Fall 2009

15

Types of Summaries (cont.)

• Output: – language chosen for summarization– format of the resulting summary

(table/paragraph/key words/documents with different sections and headings)

Page 16: I256  Applied Natural Language Processing Fall 2009

16

• Compression rate (summary length/source length)

• Audience (user-focused vs. generic)• Relation to source (extract vs. abstract)• Function (indicative vs. informative vs. critical)• Coherence: the way the parts of the text gather

together to form an integrated whole– Coherent vs. incoherent

• Incoherent: unresolved anaphors, gaps in the reasoning, sentences which repeat the same or similar meaning (redundancy) a lack of organization

Parameters

Page 17: I256  Applied Natural Language Processing Fall 2009

17

• Span (single or MDS)• Language

– monolingual, multi-lingual, cross-lingual, sub-languages (technical, tourism)

• Media• Genres

– Headlines, minutes, biographies, movie summaries, chronologies, etc.

Parameters (cont.)

Page 18: I256  Applied Natural Language Processing Fall 2009

18

Process• Three phases (typically)1. Analysis – content identification

– Analyze the input, build an internal representation of it– Can be done at different levels

• Morphology, syntax, semantics, discourse– And looking at different elements

• Sub-word, word, phrase, sentence, paragraph, document

2. Transformation (or refinement) -- conceptual organization– Transform the internal representation into a summary

representation (mostly for abstracts or MDS)3. Synthesis (Realization)

– Summary representation is rendered into natural language

Page 19: I256  Applied Natural Language Processing Fall 2009

19

Summarization approaches• Shallow approaches

– Syntactic level at most– Typically produce extracts– Extract salient parts of the source text and then

arrange and present them in some effective manner• Deeper approaches

– Sentential semantic level – Produce abstracts and the synthesis phase involves

natural language generation. – Knowledge-intensive, may require some domain

specific coding• Example: generation summaries of basketball statistics or

stock markets bulletins

Page 20: I256  Applied Natural Language Processing Fall 2009

20

Outline

• Introduction and Applications• Types of summarization tasks• Approaches and paradigms • Evaluation methods

Page 21: I256  Applied Natural Language Processing Fall 2009

21

Overview of Extraction Methods• General method:

– score each entity (sentence, word) ; combine scores; choose best sentence(s)

• Word frequencies throughout the text• Position in the text

– lead method; optimal position policy– title/heading method

• Cue phrases in sentences• Cohesion: links among words

– word co-occurrence– coreference– lexical chains

• Discourse structure of the text• Information Extraction: parsing and analysis

Page 22: I256  Applied Natural Language Processing Fall 2009

22

Using Word Frequencies

• Luhn 58: Very first work in automated summarization• Assumptions:

– Frequent words indicate the topic– Frequent means with reference to the corpus frequency– Clusters of frequent words indicate summarizing sentence

• Stemming based on similar prefix characters• Very common words and very rare words are ignored• Evaluation: straightforward approach empirically shown

to be mostly detrimental in summarization systems.

Page 23: I256  Applied Natural Language Processing Fall 2009

23

Ranked Word Frequency

Zipf’s curve

Resolving power of significant words

Page 24: I256  Applied Natural Language Processing Fall 2009

24

• Claim : Important sentences occur in specific positions– “lead-based” summary

• just take first sentence(s)!– Important information occurs in specific sections

of the document (introduction/conclusion)

• Experiments:– In 85% of 200 individual paragraphs the topic sentences

occurred in initial position and in 7% in final position

Position in the text

Page 25: I256  Applied Natural Language Processing Fall 2009

25

• Claim : title of document indicates its content – Unless editors are being cute– Not true for novels usually– What about blogs …?

• Words in title help find relevant content– Create a list of title words, remove “stop words”– Use those as keywords in order to find

important sentences

Title method

Page 26: I256  Applied Natural Language Processing Fall 2009

26

Optimum Position Policy (OPP)• Claim: Important sentences are located at positions

that are genre-dependent; these positions can be determined automatically through training– Corpus: 13000 newspaper articles (ZIFF corpus).– Step 1: For each article, determine overlap between

sentences and the index terms for the article.– Step 2: Determine a partial ordering over the locations

where sentences containing important words occur: Optimal Position Policy (OPP)

• (Some recent work looked at the use of citation sentences.)

Page 27: I256  Applied Natural Language Processing Fall 2009

27

Cue phrases method• Claim : Important sentences contain cue

words/indicative phrases– “The main aim of the present paper is to describe…”– “The purpose of this article is to review…” – “In this report, we outline…” – “Our investigation has shown that…”

• Some words are considered bonus others stigma– bonus: comparatives, superlatives, conclusive expressions,

etc.– stigma: negatives, pronouns, etc. non-important sentences

contain ‘stigma phrases’ such as hardly and impossible.

• These phrases can be detected automatically• Method: Add to sentence score if it contains a bonus

phrase, penalize if it contains a stigma phrase.

Page 28: I256  Applied Natural Language Processing Fall 2009

28

Bayesian Classifier

• Statistical learning method• Corpus

– 188 document + summary pairs from scientific journals

Page 29: I256  Applied Natural Language Processing Fall 2009

29

Bayesian Classifier• For each sentence s in the documents

– extract features• Fixed-phrase feature

– Certain phrases indicate summary, e.g. “in summary”• Paragraph feature

– Paragraph initial/final more likely to be important.• Thematic word feature

– Repetition is an indicator of importance• Uppercase word feature

– Uppercase often indicates named entities. (Taylor)• Sentence length cut-off

– Summary sentence should be > 5 words.– Calculate probability of the sentence s being in the

summary

Page 30: I256  Applied Natural Language Processing Fall 2009

30

Bayesian Classifier: Training

• Hand-label sentences in training set (good/bad summary sentences)

• Train classifier to distinguish good/bad summary sentences

• Model used: Naïve Bayes• Can rank sentences according to score

and show top n to user.

Page 31: I256  Applied Natural Language Processing Fall 2009

31

Details: Bayesian Classifier

• Assuming statistical independence:

k

j j

k

j jk

FP

SsPSsFPFFFSsP

1

121

)(

)()|(),...,|(

),()()|,...,(),...,|(

,...21

2121

k

kk FFFP

SsPSsFFFPFFFSsP

Probability that sentence s is includedin summary S, given that sentence’s feature value pairs

Probability of feature-value pairoccurring in a source sentencewhich is also in the summary

Page 32: I256  Applied Natural Language Processing Fall 2009

32

• Each Probability is calculated empirically from a corpus– See how often each feature is seen with a

sentence selected for a summary, vs. how often that feature is seen in any sentence.

• Higher probability sentences are chosen to be in the summary

• Performance:– For 25% summaries, 84% precision

Bayesian Classifier

Page 33: I256  Applied Natural Language Processing Fall 2009

33

MaxEntropy model

• Maxent model – no independence assumptions

• Features: word pairs, sentence length, sentence position, discourse features (e.g., whether sentence follows the “Introduction”, etc.)– MaxEnt outperforms Naïve Bayes

Page 34: I256  Applied Natural Language Processing Fall 2009

34

Cohesion-based methods• Claim: Important sentences/paragraphs

are the highest connected entities in more or less elaborate semantic structures.

• Classes of approaches– word co-occurrences; – local salience and grammatical relations;– co-reference;– lexical similarity (WordNet, lexical chains);– combinations of the above.

Page 35: I256  Applied Natural Language Processing Fall 2009

35

Cohesion: word co-occurrence

• Apply IR methods at the document level: texts are collections of paragraphs– Use a traditional, IR-based, word similarity

measure to determine for each paragraph Pi the set Si of paragraphs that Pi is related to.

• Method: – determine relatedness score Si for each

paragraph,– extract paragraphs with largest Si scores.

P1 P2

P3

P4

P5P6P7

P8

P9

Page 36: I256  Applied Natural Language Processing Fall 2009

36

Combining the Evidence

• Problem: which extraction methods to believe?

• Answer: assume they are independent, and combine their evidence: merge individual sentence scores.

Page 37: I256  Applied Natural Language Processing Fall 2009

37

Information extraction methods

• Idea: content selection using templates – Predefine a template, whose slots specify what

is of interest. – Use a canonical IE system to extract from a

(set of) document(s) the relevant information; fill the template.

– Generate the content of the template as the summary.

Page 38: I256  Applied Natural Language Processing Fall 2009

38

Information Extraction method

• Example template:MESSAGE:ID TSL-COL-0001SECSOURCE:SOURCE ReutersSECSOURCE:DATE 26 Feb 93

Early afternoonINCIDENT:DATE 26 Feb 93INCIDENT:LOCATION World Trade CenterINCIDENT:TYPE BombingHUM TGT:NUMBER AT LEAST 5

Page 39: I256  Applied Natural Language Processing Fall 2009

39

Information Extraction method

• Knowledge-rich approachesMESSAGE: ID TST3-MUC4-0010 MESSAGE: TEMPLATE 2 INCIDENT: DATE 30 OCT 89 INCIDENT: LOCATION EL SALVADOR INCIDENT: TYPE ATTACK INCIDENT: STAGE OF EXECUTION ACCOMPLISHED INCIDENT: INSTRUMENT ID INCIDENT: INSTRUMENT TYPEPERP: INCIDENT CATEGORY TERRORIST ACT PERP: INDIVIDUAL ID "TERRORIST" PERP: ORGANIZATION ID "THE FMLN" PERP: ORG. CONFIDENCE REPORTED: "THE FMLN" PHYS TGT: ID PHYS TGT: TYPEPHYS TGT: NUMBERPHYS TGT: FOREIGN NATIONPHYS TGT: EFFECT OF INCIDENTPHYS TGT: TOTAL NUMBERHUM TGT: NAMEHUM TGT: DESCRIPTION "1 CIVILIAN"HUM TGT: TYPE CIVILIAN: "1 CIVILIAN"HUM TGT: NUMBER 1: "1 CIVILIAN"HUM TGT: FOREIGN NATIONHUM TGT: EFFECT OF INCIDENT DEATH: "1 CIVILIAN"HUM TGT: TOTAL NUMBER

Page 40: I256  Applied Natural Language Processing Fall 2009

40

Generation

On October 30, 1989, one civilian was killed in a reported FMLN attack in El Salvador.

• Generating text from templates

Page 41: I256  Applied Natural Language Processing Fall 2009

41

Input: Cluster of templates

T1 Tm

Conceptual combiner

T2 …..

Combiner

Paragraph planner

Planningoperators

Linguistic realizer

Sentence planner

Sentence generator

Lexical chooserLexicon

OUTPUT: Base summary

SURGE

Domainontology

Page 42: I256  Applied Natural Language Processing Fall 2009

42

Operators for generation• If there are two templates

ANDthe location is the same

ANDthe time of the second template is after the time of the first template

ANDthe source of the first template is different from the source of the second template

ANDat least one slot differs

THENcombine the templates using the contradiction operator...

Page 43: I256  Applied Natural Language Processing Fall 2009

43

Operators: ContradictionContradiction

The afternoon of February 26, 1993, Reuters reported that a suspected bomb killed at least six people in the World Trade Center. However, Associated Press announced that exactly five people were killed in the blast.

Precondition:Different sources report contradictory values for a small number of slots

Other operators are: refinement, agreement…

Page 44: I256  Applied Natural Language Processing Fall 2009

44

Excerpts from four articles

JERUSALEM - A Muslim suicide bomber blew apart 18 people on a Jerusalem bus and wounded 10 in a mirror-image of an attack one week ago. The carnage could rob Israel's Prime Minister Shimon Peres of the May 29 election victory he needs to pursue Middle East peacemaking. Peres declared all-out war on Hamas but his tough talk did little to impress stunned residents of Jerusalem who said the election would turn on the issue of personal security.

JERUSALEM - A bomb at a busy Tel Aviv shopping mall killed at least 10 people and wounded 30, Israel radio said quoting police. Army radio said the blast was apparently caused by a suicide bomber. Police said there were many wounded.

A bomb blast ripped through the commercial heart of Tel Aviv Monday, killing at least 13 people and wounding more than 100. Israeli police say an Islamic suicide bomber blew himself up outside a crowded shopping mall. It was the fourth deadly bombing in Israel in nine days. The Islamic fundamentalist group Hamas claimed responsibility for the attacks, which have killed at least 54 people. Hamas is intent on stopping the Middle East peace process. President Clinton joined the voices of international condemnation after the latest attack. He said the ``forces of terror shall not triumph'' over peacemaking efforts.

TEL AVIV (Reuter) - A Muslim suicide bomber killed at least 12 people and wounded 105, including children, outside a crowded Tel Aviv shopping mall Monday, police said. Sunday, a Hamas suicide bomber killed 18 people on a Jerusalem bus. Hamas has now killed at least 56 people in four attacks in nine days. The windows of stores lining both sides of Dizengoff Street were shattered, the charred skeletons of cars lay in the street, the sidewalks were strewn with blood. The last attack on Dizengoff was in October 1994 when a Hamas suicide bomber killed 22 people on a bus.

1

2

3

4

Page 45: I256  Applied Natural Language Processing Fall 2009

45

Four templates

MESSAGE: ID TST-REU-0001 SECSOURCE: SOURCE Reuters SECSOURCE: DATE March 3, 1996 11:30 PRIMSOURCE: SOURCE INCIDENT: DATE March 3, 1996 INCIDENT: LOCATION Jerusalem INCIDENT: TYPE Bombing HUM TGT: NUMBER “killed: 18'' “wounded: 10” PERP: ORGANIZATION ID

MESSAGE: ID TST-REU-0002 SECSOURCE: SOURCE Reuters SECSOURCE: DATE March 4, 1996 07:20 PRIMSOURCE: SOURCE Israel Radio INCIDENT: DATE March 4, 1996 INCIDENT: LOCATION Tel Aviv INCIDENT: TYPE Bombing HUM TGT: NUMBER “killed: at least 10'' “wounded: more than 100” PERP: ORGANIZATION ID

MESSAGE: ID TST-REU-0003 SECSOURCE: SOURCE Reuters SECSOURCE: DATE March 4, 1996 14:20 PRIMSOURCE: SOURCE INCIDENT: DATE March 4, 1996 INCIDENT: LOCATION Tel Aviv INCIDENT: TYPE Bombing HUM TGT: NUMBER “killed: at least 13'' “wounded: more than 100” PERP: ORGANIZATION ID “Hamas”

MESSAGE: ID TST-REU-0004 SECSOURCE: SOURCE Reuters SECSOURCE: DATE March 4, 1996 14:30 PRIMSOURCE: SOURCE INCIDENT: DATE March 4, 1996 INCIDENT: LOCATION Tel Aviv INCIDENT: TYPE Bombing HUM TGT: NUMBER “killed: at least 12'' “wounded: 105” PERP: ORGANIZATION ID

43

21

Page 46: I256  Applied Natural Language Processing Fall 2009

46

Fluent summary with comparisons

Reuters reported that 18 people were killed on Sunday in a bombing in Jerusalem. The next day, a bomb in Tel Aviv killed at least 10 people and wounded 30 according to Israel radio. Reuters reported that at least 12 people were killed and 105 wounded in the second incident. Later the same day, Reuters reported that Hamas has claimed responsibility for the act.

Page 47: I256  Applied Natural Language Processing Fall 2009

47

Outline

• Introduction and Applications• Types of summarization tasks• Approaches and paradigms (for single

documents summarization)• Evaluation methods

Page 48: I256  Applied Natural Language Processing Fall 2009

48

• When a manual summary is available:

1. Choose a granularity (clause; sentence; paragraph)2. Create a similarity measure for that granularity (word

overlap; multi-word overlap, perfect match)3. Measure the similarity of each unit in the new to the

most similar unit(s)4. Measure Recall and Precision.

Evaluation methods

Page 49: I256  Applied Natural Language Processing Fall 2009

49

• When a manual summary is NOT available:

1. Intrinsic –how good is the summary as a summary?2. Extrinsic – how well does the summary help the

user?

Evaluation methods

Page 50: I256  Applied Natural Language Processing Fall 2009

50

• Intrinsic measures: how good is the summary as a summary? – Problem: how do you measure the goodness of a summary?– Studies: compare to ideal or supply criteria—fluency, quality,

informativeness, coverage, etc. • Summary evaluated on its own or comparing it with the source

– Is the text cohesive and coherent?– Does it contain the main topics of the document? – Are important topics omitted?

Intrinsic measures

Page 51: I256  Applied Natural Language Processing Fall 2009

From lecture notes by Nachum Dershowitz & Dan Cohen

Toward a Theory of Evaluation• Two Measures:

• Measuring length: – Number of letters? words?

• Measuring information: – Shannon Game: quantify information content.– Question Game: test reader’s understanding.– Classification Game: compare classifiability.

Compression Ratio: CR = (length S) / (length T)Retention Ratio: RR = (info in S) / (info in T)

Page 52: I256  Applied Natural Language Processing Fall 2009

52

• How well does the summary help a user with a task?– Problem: does summary quality correlate with

performance?– Studies: GMAT tests; news analysis; IR; text categorization

• Evaluation in an specific task – Can the summary be used instead of the document?– Can the document be classified by reading the summary?– Can we answer questions by reading the summary?

Extrinsic measures

Page 53: I256  Applied Natural Language Processing Fall 2009

53

The Document Understanding Conference (DUC)

• This is really the Text Summarization Competition– Started in 2001

• Task and Evaluation (for 2001-2004):– Various target sizes were used (10-400 words)– Both single and multiple-document summaries

assessed– Summaries were manually judged for both content

and readability. – Each peer (human or automatic) summary was

compared against a single model summary

Page 54: I256  Applied Natural Language Processing Fall 2009

54

The Document Understanding Conference (DUC)

• Made a big change in 2005– Extrinsic evaluation proposed but rejected (write a

natural disaster summary)– Instead: a complex question-focused summarization

task that required summarizers to piece together information from multiple documents to answer a question or set of questions as posed in a DUC topic.

– Also indicated a desired granularity of information

Page 55: I256  Applied Natural Language Processing Fall 2009

55

The Document Understanding Conference (DUC)

• Evaluation metrics for new task:– Grammaticality– Non-redundancy– Referential clarity– Focus– Structure and Coherence– Responsiveness (content-based evaluation)

• This was a difficult task to do well in.

Page 56: I256  Applied Natural Language Processing Fall 2009

56

DUC TAC• TAC (Text Analysis Conference)

summarization track is a continuation of DUC (since 2008) http://www.nist.gov/tac/

• Two tasks: 1. Initial: a 100-word summary of a set of 10 newswire

articles about a particular topic.2. Update: a 100-word summary of a subsequent set of

10 newswire articles for the same topic, under the assumption that the reader has already read the first 10 documents. The purpose of the update summary is to inform the reader of new information about the topic.

Page 57: I256  Applied Natural Language Processing Fall 2009

57

Available corpora & resources• DUC corpus (Document Understanding Conferences)

– http://duc.nist.gov• Text analysis conference (TAC 2009) http://www.nist.gov/tac/

– TAC 2009 Summarization Track http://www.nist.gov/tac/2009/Summarization/index.html

• SummBank corpus– http://www.summarization.com/summbank

• SUMMAC corpus– send mail to [email protected]

• <Text+Abstract+Extract> corpus– send mail to [email protected]

• Open directory project– http://dmoz.org

• www.summarization.com

Page 58: I256  Applied Natural Language Processing Fall 2009

58

Possible research topics

• Corpus creation and annotation• MMM: Multidocument, Multimedia,

Multilingual• Evolving summaries• Personalized summarization• Web-based summarization• Embedded systems• Spoken document summarization