an information-pattern-based approach to novelty detectionxli/ipm_1157.pdf · in this paper, we...

30
UNCORRECTED PROOF 1 2 An information-pattern-based approach to novelty detection 3 Xiaoyan Li a, * , W. Bruce Croft b 4 a Department of Computer Science, Mount Holyoke College, 50 College Street, South Hadley, MA 01075, United States 5 b Department of Computer Science, University of Massachusetts, Amherst, MA 01002, United States 6 Received 11 July 2007; received in revised form 19 September 2007; accepted 25 September 2007 7 8 Abstract 9 In this paper, a new novelty detection approach based on the identification of sentence level information patterns is 10 proposed. First, ‘‘novelty’’ is redefined based on the proposed information patterns, and several different types of infor- 11 mation patterns are given corresponding to different types of users’ information needs. Second, a thorough analysis of sen- 12 tence level information patterns is elaborated using data from the TREC novelty tracks, including sentence lengths, named 13 entities (NEs), and sentence level opinion patterns. Finally, a unified information-pattern-based approach to novelty detec- 14 tion (ip-BAND) is presented for both specific NE topics and more general topics. Experiments on novelty detection on 15 data from the TREC 2002, 2003 and 2004 novelty tracks show that the proposed approach significantly improves the 16 performance of novelty detection in terms of precision at top ranks. Future research directions are suggested. 17 Ó 2007 Elsevier Ltd. All rights reserved. 18 Keywords: Novelty detection; Information retrieval; Question answering; Information patterns; Named entities 19 20 1. Introduction 21 1.1. What is novelty detection? 22 Novelty detection can be viewed as going a step further than traditional document retrieval. Based on the 23 output of a document retrieval system (i.e., a ranked list of documents), a novelty detection system will further 24 extract documents with new information from the ranked list. The goal is for users to quickly get useful infor- 25 mation without going through a lot of redundant information, which is usually a tedious and time-consuming 26 task. The detection of new information is an important component in many potential applications. It is a new 27 research direction, but it has attracted increasing attention in the information retrieval field. The TREC nov- 28 elty tracks, which are related to novelty detection, were run for three years (Harman, 2002; Soboroff, 2004; 29 Soboroff & Harman, 2003). 30 Novelty detection can be performed at two different levels: the event level and the sentence level. At the 31 event level, a novel document is required to not only be relevant to a topic (i.e., a query) but also to discuss 0306-4573/$ - see front matter Ó 2007 Elsevier Ltd. All rights reserved. doi:10.1016/j.ipm.2007.09.013 * Corresponding author. Tel.: +1 413 538 2554; fax: +1 413 538 3431. E-mail address: [email protected] (X. Li). Available online at www.sciencedirect.com Information Processing and Management xxx (2007) xxx–xxx www.elsevier.com/locate/infoproman IPM 1157 No. of Pages 30, Model 3+ 24 October 2007; Disk Used ARTICLE IN PRESS Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection, Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Upload: others

Post on 27-Jul-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

1

2

3

45

67

8

9

10

11

12

13

14

15

16

17

1819

20

21

22

23

24

25

26

27

28

29

30

31

Available online at www.sciencedirect.com

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

F

Information Processing and Management xxx (2007) xxx–xxx

www.elsevier.com/locate/infoproman

RO

OAn information-pattern-based approach to novelty detection

Xiaoyan Li a,*, W. Bruce Croft b

a Department of Computer Science, Mount Holyoke College, 50 College Street, South Hadley, MA 01075, United Statesb Department of Computer Science, University of Massachusetts, Amherst, MA 01002, United States

Received 11 July 2007; received in revised form 19 September 2007; accepted 25 September 2007

P

EC

TEDAbstract

In this paper, a new novelty detection approach based on the identification of sentence level information patterns isproposed. First, ‘‘novelty’’ is redefined based on the proposed information patterns, and several different types of infor-mation patterns are given corresponding to different types of users’ information needs. Second, a thorough analysis of sen-tence level information patterns is elaborated using data from the TREC novelty tracks, including sentence lengths, namedentities (NEs), and sentence level opinion patterns. Finally, a unified information-pattern-based approach to novelty detec-tion (ip-BAND) is presented for both specific NE topics and more general topics. Experiments on novelty detection ondata from the TREC 2002, 2003 and 2004 novelty tracks show that the proposed approach significantly improves theperformance of novelty detection in terms of precision at top ranks. Future research directions are suggested.� 2007 Elsevier Ltd. All rights reserved.

Keywords: Novelty detection; Information retrieval; Question answering; Information patterns; Named entities

UN

CO

RR

1. Introduction

1.1. What is novelty detection?

Novelty detection can be viewed as going a step further than traditional document retrieval. Based on theoutput of a document retrieval system (i.e., a ranked list of documents), a novelty detection system will furtherextract documents with new information from the ranked list. The goal is for users to quickly get useful infor-mation without going through a lot of redundant information, which is usually a tedious and time-consumingtask. The detection of new information is an important component in many potential applications. It is a newresearch direction, but it has attracted increasing attention in the information retrieval field. The TREC nov-elty tracks, which are related to novelty detection, were run for three years (Harman, 2002; Soboroff, 2004;Soboroff & Harman, 2003).

Novelty detection can be performed at two different levels: the event level and the sentence level. At theevent level, a novel document is required to not only be relevant to a topic (i.e., a query) but also to discuss

0306-4573/$ - see front matter � 2007 Elsevier Ltd. All rights reserved.

doi:10.1016/j.ipm.2007.09.013

* Corresponding author. Tel.: +1 413 538 2554; fax: +1 413 538 3431.E-mail address: [email protected] (X. Li).

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 2: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

32

33

34

35

36

37

38 Q2

39

40

41

42

43

44

45

46

47

48

49

50

51

52

53

54

55

56

57

58

59

60

61

62

63

64

65

66

67

68

69

70

71

72

73

74

75

76

77

78

79

80

81

2 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

CO

RR

EC

TED

PR

OO

F

a new event. At the sentence level, a novel sentence should be relevant to a topic and provide new information.This means that a novel sentence may either discuss a new event or provide new information about an oldevent. Novelty detection at the sentence level is also the basis for novelty detection at the event level. Thereforethe work in this paper will focus on novelty detection at the sentence level.

1.2. Motivation and contribution of the work

A variety of novelty measures have been described in the literature (Allan, Wade, & Bolivar, 2003; Harman,2002; Zhang, Callan, & Minka, 2003; Zhang, Lin, Liu, Zhao, & Ma, 2002; Zhang et al., 2003). The variousdefinitions of novelty, however, are quite vague and seem only indirectly related to intuitive notions of novelty.For example, in Zhang, Callan et al. (2002) and Soboroff (2003), novelty detection was informally defined asthe opposite of redundancy. In Allan et al. (2003), it was described as new information based on new wordsexisting in a sentence. In Harman (2002) and Soboroff (2003), no clear definition of novelty was provided.Usually, new words appearing in an incoming sentence or document contribute to the novelty scores in var-ious novelty measures in different ways. However, new words are not equivalent to novelty (new information).For example, rephrasing a sentence with a different vocabulary does not mean that this revised sentence con-tains new information that is not covered by the original sentence. Furthermore, new words appearing inmaterial that is not relevant to a user’s query do not provide any new information with respect to the user’sinterest, even though it may contain some new information useful to other users. A more precise definition ofnovelty or new information is required.

Let us look at a real example to understand the limitations of the related work and why we are proposing anew approach for novelty detection. Topic (query) 306 from the TREC novelty track 2002 is about ‘‘AfricanCivilian Deaths’’. The user is asking for the number of civilian non-combatants that have been killed in thevarious civil wars in Africa. Therefore, a number should appear in sentences that are relevant to the query.Let us consider the following four sentences given in Example 1.

Example 1. Topic: ‘‘African Civilian Deaths’’

ÆNumberæ: 306ÆTitleæ African Civilian DeathsÆDescriptionæ: How many civilian non-combatants have been killed in the various civil wars in Africa?ÆNarrativeæ: A relevant document will contain specific casualty information for a given area, country, orregion. It will cite numbers of civilian deaths caused directly or indirectly by armed conflict.Sentence 1 (Relevant): ‘‘It could not verify Somali claims of more than 100 civilian deaths’’.Sentence 2 (Relevant): ‘‘Natal’s death toll includes another massacre of 11 ANC [African National Con-gress] supporters’’.Sentence 3 (Non-relevant): ‘‘Once the slaughter began, following the death of President Juvenal Habyari-mana in an air crash on April 6, hand grenades were thrown into schools and churches that had given ref-uge to Tutsi civilians’’.Sentence 4 (Non-relevant): ‘‘A Ghana News Agency correspondent with the West African force said thatrebels loyal to Charles Taylor began attacking the civilians shortly after the peace force arrived in Monrovialast Saturday to try to end the eight-month-old civil war.’’

UNEach of the four sentences has two query terms (in italics) that match the key words from the query. How-

ever, only the first two sentences are relevant sentences. In addition to the two matching words, each of thefirst two sentences also has a number, 100 and 11 (in bold and underlined), respectively. Hence, the firsttwo sentences are both topically relevant to the query and have the right type of information with respectto the user’s request behind the query. The third sentence and the fourth sentence are not relevant to the querymainly because they do not contain numbers that are required by the user. However, for the example querygiven above, it is very difficult for traditional word-based approaches to separate the two non-relevantsentences (sentence 3 and sentence 4) from the two relevant sentences (sentence 1 and sentence 2). Even worse,the two non-relevant sentences are very likely to be identified as novel sentences simply because they containmany new words that do not appear in previous sentences.

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 3: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

82

83

84

85

86

87

88

89

90

91

92

93

94

95

96

97

98

99

100

101

102

103

104

105

106

107

108

109

110

111

112

113

114

115

116

117

118

119

120

121

122

123

124

125

126

127

128

129

130

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 3

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

TED

PR

OO

F

To solve this problem, a deeper query understanding beyond matching key words and a more precise nov-elty understanding beyond identifying new words is required. This motivated our work on information-pat-tern-based novelty detection. In this paper, we investigate if identifying query-related named entity (NEs)patterns in sentences will significantly improve the performance in novelty detection, particularly at top ranks(Li & Croft, 2005). The basic concept of our NE-pattern-based approach is that each query could be treated asmultiple specific NE-questions. Each question is represented by a few query words, and it requires a certaintype of named entities as answers.

However, queries (or topics) that can be transformed into specific NE-questions are only a small portion ofthe possible queries. For example, in TREC 2003, there are only 15 (out of 50) topics that can be formulatedinto specific NE-questions. The rest of the topics will be called general topics throughout the paper, since theycan only be formulated into general questions. New and effective information patterns are needed in order tosignificantly improve the performance of novelty detection for those general topics. As one of the most inter-esting sentence level information patterns, we have found that the detection of information patterns related toopinions, i.e., opinion patterns, is very effective in improving the performance of the general topics (Li & Croft,2006). As an example, Topic N1, from the TREC novelty track 2003, is about ‘‘partial birth abortion ban’’.This is a query that cannot be easily converted into any specific NE-questions. However, we know that theuser is trying to find opinions about the proposed ban on partial birth abortions. Therefore, relevant sentencesare more likely to be ‘‘opinion sentences’’. Let us consider the following two sentences.

Example 2. Topic: ‘‘Partial Birth Abortion Ban’’

ÆNumberæ: N1ÆTitleæ: partial birth abortion banÆToptypeæ: opinionÆDescriptionæ: Find opinions about the proposed ban on partial birth abortions.ÆNarrativeæ: Relevant information includes opinions on partial birth abortion and whether or not it shouldbe legal. Opinions that also cover abortion in general are relevant. Opinions on the implications of pro-posed bans on partial birth abortions and the positions of the courts are also relevant.Sentence 1 (Relevant and Novel): ‘‘The court’s ruling confirms that the entire campaign to ban ‘partial birth

abortion’ – a campaign that has consumed Congress and the federal courts for over three years – is nothingbut a fraud designed to rob American women of their right to abortion,’’ said Janet Benshoof, president ofCenter for Reproductive Law and Policy.Sentence 2 (Non-relevant): Since the Senate’s last partial birth vote, there have been 11 court decisions onthe legal merits of partial birth bans passed by different states.

Both sentence 1 and sentence 2 have five matched terms (in italics). But only sentence 1 is relevant to thetopic. Note that in addition to the matched terms, sentence 1 also has opinion patterns, indicated by the word‘‘said’’ and a pair of quotation marks (both in bold and underlined). The topic is an opinion topic that requiresrelevant sentences to be opinion sentences. The first sentence is relevant to the query because it is an opinionsentence and therefore topically related to the query. However, for the topic given in the above example, it isvery difficult for traditional word-based approaches to separate the non-relevant sentence (sentence 2) fromthe relevant sentence (sentence 1). This work tries to attack this hard problem.

For a useful novelty detection system, a unified framework of the pattern-based approach is also requiredto deal with both the specific and the general topics. We propose a unified information-pattern-basedapproach to novelty detention (ip-BAND), which includes the following three steps: query analysis, relevantsentence retrieval and new pattern detection.

In summary, this work makes the following three contributions:

(1) We provide a new and more explicit definition of novelty. Novelty is defined as new answers to the poten-

tial questions representing a user’s request or information need.(2) We propose a new concept in novelty detection – query-related information patterns. Very effective infor-

mation patterns for novelty detection at the sentence level have been identified.

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 4: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

131

132

133

134

135

136

137

138

139

140

141

142

143

144

145

146

147

148

149

150

151

152

153

154

155

156

157

158

159

160

161

162

163

164

165

166

167

168

169

170

171

172

173

174

4 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

PR

OO

F

(3) We propose a unified pattern-based approach that includes the following three steps: query analysis, rel-evant sentence detection and new pattern detection. The unified approach works for both specific topicsand general topics.

1.3. Organization of the paper

The rest of the work is organized as follows. Section 2 gives a review of related work on novelty detection.Section 3 describes data collections from TREC novelty tracks (2002–2004). Evaluation measures used inTREC and the ones we are using are also discussed in this section. Section 4 introduces our new definitionof ‘‘novelty’’, and the concept of the proposed information patterns for novelty detection of both specificand general topics. Section 5 gives a thorough analysis of sentence level information patterns including sen-tence lengths, named entities, and opinion patterns. The analysis is performed on data from the TREC2002 and/or 2003 novelty tracks, which provides guidelines in applying those patterns in novelty detection.Section 6 describes the proposed unified pattern-based approach to novelty detection for both general and spe-cific topics. Different treatments for specific topics and general topics will be highlighted. Section 7 showsexperimental results in using the proposed information patterns for significantly improving the performanceof novelty detection for specific topics, general topics, and the overall performance of novelty detection for alltopics using the unified approach. Section 8 summarizes the work and indicates some future researchdirections.

D

CTE2. Related work

In the following, we give a review of research related to novelty detection at the two different levels, theevent level and the sentence level, as well in other applications. Then, we point out the main differencesbetween the pattern-based approach proposed in this paper and other approaches in the literature.

UN

CO

RR

E2.1. Novelty detection at the event level

Work on novelty detection at the event level arises from the Topic Detection and Tracking (TDT)research, which is concerned with online new event detection and/or first story detection (Allan, Lavrenko,& Jin, 2000; Allan, Paka, & Lavrenko, 1998; Brants, Chen, & Farahat, 2003; Franz, Ittycheriah, McCarley,& Ward, 2001; Kumaran & Allan, 2004; Stokes & Carthy, 2001; Yang, Pierce, & Carbonell, 1998; Yang,Zhang, & Jin, 2002). Current techniques for new event detection are usually based on clustering algorithms,which include two important issues: event modeling and event clustering. Several models, such as vector spacemodels, language models, lexical chains, etc., have been proposed to represent incoming new stories/docu-ments. Each story is then grouped into clusters. An incoming story will either be grouped into the closestcluster if the similarity score between them is above a preset similarity threshold or it will be used to starta new cluster. A story which starts a new cluster will be marked as the first story about a new topic, or itwill be marked as ‘‘old’’ (about an old event) if there exists a novelty threshold (which is smaller than thesimilarity threshold) and the similarity score between the story and its closest cluster is greater than the nov-elty threshold.

Temporal proximity has also been considered, e.g., in Yang et al. (1998) and Franz et al. (2001), based onthe observation that news stories discussing the same event tend to be temporally proximate. Yang et al. (2002)proposed a two-level scheme first story detection system. The limitation of this technique is that it needs topredefine the taxonomy of broad topics and to have training data for each broad topic, which makes it inap-plicable in the TDT context. Kumaran and Allan (2004) proposed a multi-stage new event detection (NED)system. Stories were classified into categories and NED was performed within categories. Stokes and Carthy(2001) proposed a composite document representation, which combined a free text vector and a lexical chaincreated using WordNet, and achieved a marginal increase in effectiveness.

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 5: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

175

176

177

178

179

180

181

182

183

184

185

186

187

188

189

190

191

192

193

194

195

196

197

198

199

200

201

202

203

204

205

206

207

208

209

210

211

212

213

214

215

216

217

218

219

220

221

222

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 5

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

OR

REC

TED

PR

OO

F

2.2. Novelty detection at the sentence level

Research on novelty detection at the sentence level is related to the TREC novelty tracks described in moredetail in the next section. The goal is to find relevant and novel sentences, given a query and an ordered list ofrelevant documents. Many research groups participated in the three TREC novelty tracks and reported theirresults (Allan et al., 2003; Eichmann & Srinivasan, 2002; Jin, Zhao, & Xu, 2003; Kazawa, Hirao, Isozaki, &Maeda, 2003; Kwok, Deng, Dinstl, & Chan, 2002; Li, 2003; Qi, Otterbacher, Winkel, & Radev, 2002; Sunet al., 2003; Tsai, Hsu, & Chen, 2003; Zhang, Callan et al., 2002; Zhang et al., 2002, 2003).

Novelty detection at the sentence level can usually be conducted in two steps: relevant sentence retrievaland novel sentence extraction. In current techniques developed for novelty detection at the sentence level,new words appearing in ‘‘relevant’’ sentences usually contribute to the scores that are used to rank sentencesfor novelty. Many similarity functions used in information retrieval are also used in novelty detection. Usuallya high similarity score between a sentence and a given query will increase the relevance rank of the sentence,whereas a high similarity score between the sentence and all previously seen sentences will decrease the noveltyranking of the sentence.

The simplest novelty measure, New Word Count Measure(Allan et al., 2003), simply counts the number ofnew words appearing in a sentence and takes it as the novelty score for the sentence. There are other similarnovelty or redundancy measures that consider new words appearing in sentences. One such novelty measure,called New Information Degree (NID), was defined by Jin et al. (2003). Zhang et al. (2003) used an overlap-based redundancy measure in their work on novelty detection. The redundancy score of a sentence againstanother sentence is the number of matching terms normalized by the length of the sentence. They set theredundancy threshold to 0.55. Sentences with redundancy scores higher than the threshold were treated asredundant sentences and thus eliminated. Instead of counting new words appearing in sentences, Litkowski(2003) checked all discourse entities in a sentence. Discourse entities are semantic objects and they can havemultiple syntactic realizations within a text. A pronoun, a noun phrase and a person’s name could all belongto one single entity and they would be treated as the same discourse entity. The sentence was accepted as novelif it had new discourse entities that were not found in the growing history list of discourse entities. Eichmannand Srinivasan (2002) considered the number of new named entities and noun phrases appearing in a sentence.A sentence was declared novel if the number of new named entities and noun phrases is above a pre-declarednumber.

Many similarity functions (such as cosine similarity, language model measures etc.) used in informationretrieval have also been tried in novelty detection. Tsai et al. (2003) represented sentences as vectors and com-puted similarity scores for a sentence currently being considered with each previous sentence with the cosinesimilarity function for detecting novel sentences. The Maximal Marginal Relevance model (MMR), intro-duced by Carbonell and Goldstein (1998), was used by Sun et al. (2003) in their work on novelty detection.

Most novelty detection methods proposed in the past are highly sensitive to the quality of the initial sen-tence retrieval phase. Therefore, improving sentence retrieval performance is one of the major focuses of thispaper.

UN

C

2.3. Novelty detection in other applications

Novelty detection can be also seen in other applications, such as temporal summaries of news topics, doc-ument filtering and minimal document set retrieval, where new information detection is an important compo-nent. Those applications could include novelty detection at the event level, novelty detection at the sentencelevel, or a combination of both. In the work by Allan, Gupta, and Khandelwal (2001) on temporal summariesof new topics, the task is to extract a single sentence from each event within a news topic, where the stories arepresented one at a time and sentences from a story must be ranked before the next story can be considered. Inthe work by Zhang, Callan et al., 2002 on novelty and redundancy detection in adaptive filtering, the goal is toeliminate redundant documents. A redundant document is defined as the one that is relevant to the user’s pro-file, but only contains information that is covered by previous documents. Other research that performs nov-elty detection includes subtopic retrieval (Zhai, Cohen, & Lafferty, 2003). The goal of subtopic retrieval is to

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 6: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

223

224

225

226

227

228

229

230

231

232

233

234

235

236

237

238

239

240

241

242

243

244

245

246

247

248

249

250

251

252

253

254

255

256

257

258

6 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

find documents that cover as many different subtopics as possible. In the recent work by Dai and Srihari(2005) on minimal document set retrieval, the authors tried three different approaches to retrieve and rankdocument sets with maximum coverage but minimal redundancy of subtopics in each set: a novelty-basedalgorithm, a cluster-based generation algorithm, and a subtopic extraction based algorithm.

CTE

DPR

OO

F2.4. Comparison of our approach and others

There are three main differences between our pattern-based approach and the aforementioned approachesin the literature. First, none of the work described above gives an explicit definition of novelty, which webelieve is essential in novelty detection. Usually new words, phrases or named entities appearing in incomingsentences or documents contribute to the novelty score, though in different ways. Our pattern-based approachis based on the new definition of novelty introduced in this paper, which treats new information as new answersto questions that represented users’ information requests.

Second, in the systems that are related to the TREC novelty tracks, either the title query or all the threesections of a query were used merely as a bag of words, whereas we try to understand a query beyond bag-of-words by transforming it into multiple specific NE-questions or a general question. We believe that a user’sinformation request can be better captured with questions than a few keywords, and a deeper understandingof the user’s request helps a novelty detection system to serve the user better.

Third, the proposed information-pattern-based approach is inspired by question answering techniques andis similar to passage retrieval for factoid questions. In our approach, each query could be treated as multiplequestions; each question is represented by a few query words, and it requires a certain type of named entities asanswers. Answer sentences are defined as sentences with answers to the multiple questions representing auser’s information request. Novel sentences are sentences with new answers to those questions. Instead ofexplicitly extracting exact answers as in factoid question answering systems, we propose to first retrieve answersentences with certain information patterns that include both query words and required answer types, indicat-ing the presence of potential answers to the questions, and then identify novel sentences that are more likely tohave new answers to the questions.

NC

OR

RE

3. TREC novelty track data and evaluation

3.1. Data collections

Currently, there are three data collections from the TREC 2002, 2003 and 2004 novelty tracks officiallyavailable for experiments on novelty detection at the sentence level. Table 3.1 summarizes the statistics ofthe three data collections. There are three main differences among the three sets.

(1) The TREC 2003 and 2004 novelty track collections exhibited greater redundancy than the TREC 2002and thus has less novel sentences. Only 41.4% and 65.7% of the total relevant sentences were markednovel for the TREC 2004 novelty track and the TREC 2003 novelty track, respectively, whereas90.9% of the total relevant sentences in the 2002 track are novel sentences.

(2) TREC 2003 and TREC 2004 topics were classified into event topics and opinion topics. This allowedthese two types of topics to be treated differently for performance improvement.

U

Table 3.1Statistics of the collections from TREC 2002, 2003 and 2004 novelty tracks

Datacollections

# oftopics

# of eventtopics

# of opiniontopics

# ofsentences

# of relevantsentences

# of novelsentences

# of non-relevantsentences

TREC2002 49 N/A N/A 57227 1365 1241 55762TREC2003 50 28 25 39820 15557 10226 24263TREC2004 50 25 25 52447 8343 3454 44104

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 7: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

259

260

261

262

263

264

265

266

267

268

269

270

271

272

273

274

275

276

277

278

279

280

281

282

283

284

285

286

287

288

289

290

291

292

293

294

295

296

297

298

299

300

301

302

303

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 7

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

F

(3) TREC 2002 and 2004 novelty track collections exhibited greater difficulty in terms of relevant sentenceretrieval because there were only a small portion of sentences in the document set marked relevantsentences.

All three data sets are used in the experiments of this paper. The datasets from TREC 2002 and 2003 areused for training our system and the TREC 2004 dataset is set aside for testing.

NC

OR

REC

TED

PR

OO3.2. Evaluation issues

The TREC novelty tracks used three evaluation measures for a novelty detection system: set precision,set recall and the F measure. One problem with set recall and set precision is that they do not average wellwhen the sizes of assessors’ sets vary widely across topics. Harman (2002) illustrated the problem with anexample, where a system did precisely the wrong thing but got an average score of 0.5 for both recall andprecision. The average of the F measure is more meaningful than the average of set precision or recall whenthe sizes of the judgment sets vary widely. However, systems with an equal value of the F measures canhave a range of precision and recall scores. Hence, the F measure is not very precise in characterizing users’real requirements, such as a quick search on the Web for obtaining useful information within the topranked results.

Usually, the number of relevant (or novel) sentences retrieved by the system is determined by their rankingscores. Those sentences whose ranking scores are larger than a threshold are returned. This indicates thatthe numbers of sentences for different queries may vary dramatically, and the difference between numbersof sentences for two topics can be more than a hundred. However, in real applications, a user may only careabout how much information within a limited number of sentences, say 10 or 20, retrieved by a system. Forinstance, many users who search information on the Web only look at top 10 results in the first page returnedfrom a search engine. Few users would read the results beyond the first two pages.

Similarly, the pattern-based approach we propose is a precision-oriented approach designed to meet theaforementioned requirements from users. In improving the precision of relevance and novelty detection, somerelevant and novel sentences are filtered out. For example, in dealing with topics that can be turned into multi-ple specific questions, only those sentences that ‘‘answer’’ these identified specific questions are selected as rel-evant; other relevant sentences that are not covered by these specific questions are filtered out. In this sense,the task of our pattern-based approach, which is more useful in practice, is different from the tasks defined inthe TREC novelty tracks.

Since we apply very strict rules in the selection process, ideally, those sentences that are retrieved by oursystem are truly relevant sentences or novel sentences, and therefore the precision of the relevant and novelsentence detection should be high, particularly within top ranked results. However, the recall might be lowfor the same reason. Hence, it is not very meaningful to use set precision, set recall or the F measure to com-pare our approach with the existing methods. Instead, in this paper, we use the precision of relevant or novelsentences at top ranks. We believe that precision values at top ranks (top 5, 10, 15, 20 or 30 sentences) aremore useful in real applications where users want to quickly get some required information without goingthrough a lot of non-relevant as well as redundant information. Nevertheless, we thoroughly compare the per-formance of our system with several well-designed baselines, which have been commonly used in noveltydetection, using the same evaluation measures.

U

4. Definitions of novelty and information patterns

4.1. Our novelty definition

The definition of novelty or ‘‘new’’ information is crucial for the performance of a novelty detection system.However, as we have pointed out, novelty is usually not clearly defined in the literature. Therefore, we give ourdefinition of novelty as follows:

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 8: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

304

305

306

307

308

309

310

311

312

313

314

315

316

317

318

319

320

321

322

323

324

325

326

327

328

329

330

331

332

333

334

335

336

337

338

339

340

341

TableWord

Catego

Name

TimeNumb

8 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

PleaInfo

Novelty or new information means new answers to the potential questions representing a user’s request or

information need.

PR

OO

F

There are two important aspects in this definition. First, a user’s query will be transformed into one or morepotential questions for identifying corresponding query-related information patterns, which include both querywords and required answer types. Second, new information is obtained by detecting those sentences thatinclude previously unseen ‘‘answers’’ corresponding to the query-related patterns. We emphasize thatalthough a user’s information need is typically represented as a query consisting of a few keywords, our obser-vation is that a user’s information need may be better captured by one or more questions that lead to corre-sponding information patterns.

This novelty definition is a general one that works for novelty detection with any query that can beturned into questions. In this paper, we show that any query (topic) in the TREC novelty tracks can beturned into either one or more specific NE-questions, or a general question. Specific topics (correspondingto specific NE-questions) are those topics whose answers are specific named entities (NEs), including per-sons, locations, dates, time, numbers, etc. (Bikel, Schwartz, & Weischedel, 1999). General topics (corre-sponding to general questions), on the other hand, require obtaining additional information patterns foreffective novelty detection. Other types of questions may be further explored within the framework of thisnovelty detection.

REC

TED4.2. Information patterns

4.2.1. Information patterns for specific topics

The identification and extraction of information patterns is crucial in our approach. The information pat-tern corresponding to a specific NE-question (generated from a specific topic/query) is represented by both thequery words (of the potential question) and an NE answer type (which requires named entities as its potentialanswer). They are called NE words patterns, related to the questions about DATE (‘‘when’’), LOCATION(‘‘where’’), PERSON (‘‘who’’), ORGANIZATION (‘‘what/who’’) and NUMBER (‘‘how many’’).

For each type of the five NE-questions, a number of word patterns were constructed for question type iden-tification. Typical NE patterns are listed in Table 4.1; some were extracted from the TREC 2002 novelty trackqueries manually and some were selected from a question answering system (Li, 2003). Each NE word patternis a combination of both query words (of potential questions) and answer types (which requires named entitiesas potential answers).

NC

OR4.2.2. Information patterns for general topics

For a general topic, it is very difficult (if not impossible) to identify a particular type of named entity as itsanswer. Any type of named entity or even single words or phrases could be part of an answer as long as theanswer context is related to the question. Further, in many relevant and novel sentences, no named entities areincluded. This observation is supported by the data analysis in Section 5. Simply using named entities seemsnot very helpful for improving the performance for general topics, even though the performance of noveltydetection for specific topics is significantly improved with the use of NE patterns (Li & Croft, 2005). There-fore, the challenging questions are how to effectively make use of these named entities, and what kinds of addi-tional and critical information patterns will be effective for general topics.

U4.1patterns for the five types of NE-questions

ries Answer types Word patterns

Person who, individual, person, people, participant, candidate, customer, victim, leader, member, player, nameOrganization who, company, companies, organization, agency, agencies, name, participantLocation where, location, nation, country, countries, city, cities, town, area, regionDate when, date, time, which year, which month, which day

er Number how many, how much, length, number, polls, death tolls, injuries, how long,

se cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,rmation Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 9: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

342

343

344

345

346

347

348

349

350

351

352

353

354

355

356

357

358

359

360

361

362

363

364

365

366

367

368

369

370

371

372

373

374

375

376

377

378

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 9

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

After analyzing the TREC data, we have found that the following three kinds of information patterns arevery effective for this purpose: sentence lengths, named entity combinations and opinion patterns. We have alsofound that these patterns are effective for both general and specific topics. Details will be provided in Section 5;in the following, we introduce a particularly effective type of information patterns – opinion patterns.

ED

PR

OO

F4.2.3. Opinion patterns and opinion sentences

We note that the topics in TREC 2003 and 2004 novelty tracks are either classified as event topics or opin-ion topics. According to Soboroff (2004), ‘‘identifying sentences that contain an opinion remained a hardproblem. Event topics proved to be easier than opinion topics.’’ As a specific interesting finding, we havefound that a large portion of the general questions are about opinions. Opinions can typically be identifiedby looking at such sentence patterns as ‘‘XXX said’’, ‘‘YYY reported’’, or as marked by a pair of quotationmarks. Currently, we have identified about 20 such opinion-related sentence patterns. The full list of opinionpatterns is described in Section 5. These patterns are extracted manually from a training set that is composedof about 100 documents from the TREC 2003 novelty track. The opinion patterns currently used in our sys-tem are individual words or a sequence of words, such as said, say, claim, agree, found that, state that, etc.Note that the terms remain in their original verbal forms without word stemming, in order to more preciselycapture the real opinion sentences. For example, a word ‘‘state’’ does not necessarily indicate an opinion pat-tern, but the word combination ‘‘stated that’’ most probably does. If a sentence includes one or more opinionpatterns, it is said to be an opinion sentence.

Note that the above information patterns could also be usable to other collections. However, naturallanguage processing (NLP) is a complicated language understanding problem, and our NLP-like approachonly provides the first step to extracting suitable information patterns for novelty detection. Further workis needed to automatically learn and to extract suitable information patterns for adapting to different datacollections.

T

REC5. Information pattern analysis

Novelty detection includes two consecutive steps: first retrieving relevant sentences and then detecting novelsentences. In this section, we perform a statistical analysis of information patterns (or features) in relevant sen-tences, novel sentences and non-relevant sentences. The three information patterns studied here are: sentencelengths, named entities, and opinion patterns. The goal is to discover effective ways to use these informationpatterns in distinguishing relevant sentences from non-relevant ones, and novel sentences from non-novel ones.

NC

OR5.1. Statistics on sentence lengths

The first type of pattern we have studied is the length of a sentence. This is probably the simplest feature ofa sentence. The hypothesis is that it may include some information about the importance of the sentence. Thestatistics of sentence lengths in TREC 2002 and 2003 datasets are shown in Table 5.1. The length of a sentenceis measured in the number of words after stop words are removed from the sentence. Interestingly, we havefound that the average lengths of relevant sentences from the 2002 data and the 2003 data are 15.58 and 13.1,respectively, whereas the average lengths of non-relevant sentences from the 2002 data and the 2003 data areonly 9.55 and 8.5, respectively.

UTable 5.1Statistics of sentence lengths

Types of sentences (S) TREC 2002: 49 topics TREC 2003: 50 topics

# of S Length # of S Length

Relevant 1365 15.58 15557 13.1Novel 1241 15.64 10226 13.3Non-relevant 55862 9.55 24263 8.5

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 10: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

379

380

381

382

383

384

385

386

387

388

389

390

391

392

393

394

395

396

397

398

399

400

401

402

403

404

405

406

407

408

409

410

411

412

10 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

DPR

OO

F

To conclude, here is our first observation on sentence length (SL) patterns:

SL Observation #1. Relevant sentences on average have more words than non-relevant sentences.

The sentence length feature is quite simple, but very effective since the length differences between non-rel-evant and relevant sentences are significant. This feature is ignored in other approaches, mainly because, in thepast, sentence retrieval was performed with information retrieval techniques developed for document retrieval,where document lengths were usually used as a penalty factor. A long document may discuss multiple topicsand a short document may focus on one topic. Therefore, in document retrieval, a short document is usuallyranked higher than a long document if the two documents have same occurrences of query words. But at thesentence level, it turns out that relevant sentences have more words than non-relevant sentence on average.Therefore, this observation will be incorporated into the retrieval step to improve the performance of relevantsentence retrieval, which will then boost the performance for identifying novel sentences. Further, we have asecond observation on sentence length patterns from the statistics:

SL Observation #2: The difference in sentence lengths between novel and non-relevant sentences is slightly lar-

ger than the difference in sentence lengths between relevant sentences and non-relevant sentences.

This indicates that the incorporation of the sentence length information in relevance ranking will put thenovel sentences at higher ranks in the relevance retrieval step, which increases their chance to be selected asnovel sentences.

OR

REC

TE5.2. Statistics on opinion patterns

Topics in TREC 2003 and 2004 novelty track data collections are classified into event and opinion topics.There are 22 opinion topics out of the 50 topics from the 2003 novelty track. The number is 25 out of 50 forthe 2004 novelty track. There are no classifications of opinion and event topics in the 2002 novelty track. Wedesignate a sentence to be an opinion sentence if it has one or more opinion patterns. The list of opinion pat-terns we here identified is shown in Table 5.2. The list is by no means complete; however we want to show thatby making use of this piece of information, the performance of novelty detection can be improved. Intuitively,for an opinion topic, opinion sentences are more likely to be relevant sentences than non-opinion sentences.This hypothesis is tested with a data analysis on the 22 opinion topics from the 2003 novelty track.

We have run our statistical analysis of opinion patterns on the 22 opinion topics from the 2003 noveltytrack in order to obtain guidelines for using opinion patterns for both 2003 and 2004 data. According tothe results shown in Table 5.3, 48.1% of relevant sentences and 48.6% of the novel sentences are opinion sen-tences, but only 28.4% of non-relevant sentences are opinion sentences. This could have a significant impact onseparating relevant and novel sentences from non-relevant sentences for opinion topics, therefore we summa-rize these statistical results as Opinion Pattern (OP) Observations #1 and #2:

UN

C

Table 5.2Examples of opinion patterns

‘‘ ’’, said, say, according to, add, addressed, agree, agreed, disagreed, argue, affirmed, reaffirmed, believe, believes, claim, concern,consider, expressed, finds that, found that, fear that, idea that, insist, maintains that, predicted, reported, report, state that, statedthat, states that, show that, showed that, shows that, think, wrote

Table 5.3Statistics on opinion patterns for 22 opinion topics (2003)

Sentences (S) Total # of S # of opinion S (and %)

Relevant 7755 3733 (48.1%)Novel 5374 2609 (48.6%)Non-relevant 13360 3788 (28.4%)

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 11: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

413

414

415

416

417

418

419

420

421

422

423

424

425

426

427

428

429

430

431

432

433

434

435

436

437

438

439

440

441

443443

444

445

446

447

448

449

450

451

452

453

454

455

456

457

458

459

460

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 11

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

PR

OO

F

OP Observation #1: There are relatively more opinion sentences in relevant (and novel) sentences than in non-

relevant sentences.

OP Observation #2: The difference of numbers of opinion sentences in novel and non-relevant sentences is

slightly larger than that in relevant and non-relevant sentences.

Note that the number of opinion sentences in the statistics only counts those sentences that have one ormore opinion patterns shown in Table 5.2. Some related work (Kim, Ravichandran, & Hovy, 2004) has beendone very recently in classifying words into opinion-bearing words and non-opinion-bearing words, usinginformation from several major sources such as WordNet, World Street Journal, and the General InquirerDictionary. Using opinion-bearing words may cover more opinion sentences, but the accuracy of classifyingopinion words is still an issue. We believe that a more accurate classification of opinion sentences based on theintegration of the results of that work into our framework could further enlarge the difference in numbers ofopinion sentences between relevant sentences and non-relevant sentences. The second observation indicatesthat novel sentences may benefit more from the use of opinion patterns.

The division of opinion and event topics is given by the TREC task. In realistic situations, this informationwould not be available. However, it is possible to automatically classify the queries into these categories, par-ticularly when a query with more information than merely a few keywords is provided by a user.

EC

TED5.3. Statistics on named entities

As we have pointed out, answers and new answers to specific NE-questions are named entities. And formany of the general topics (questions), named entities are also major parts of their answers. Therefore, under-standing the distribution of named entity patterns could be very helpful both in finding relevant sentences andin detecting novel sentences.

The statistics of all the 21 named entities that can be identified by our system are listed in Table 5.4 onTREC 2002 and TREC 2003 novelty track data collections. These named entities are also grouped into fourcategories: Name, Time, Number and Object. In the table, each item lists two measurements: NY(X), andPY(X). The former is the number of Y-type (Y is either R for relevant or NR for non-relevant) sentences, eachof which has at least one of X-type named entity (X is one of the 21 NE types). The latter is the percentage ofthose sentences among all the relevant (or non-relevant) sentences in each collection. These two measurementsare calculated as

PleaInfo

R

NY ðX Þ ¼ # of Y -type sentences with at least one X -type NE ð5:1ÞP Y ðX Þ ¼ N Y ðX Þ=MY ð5:2Þ

R

UN

COIn Eq. (5.2), MY denotes the total number of Y-type (either R for relevant or NR for non-relevant) sentences.

Those numbers are listed in the head of the table. From these statistics, we have found that the five most fre-quent types of NEs are PERSON, ORGANIZATION, LOCATION, DATE and NUMBER. For each ofthem, more than 25% relevant sentences have at least one named entity of the type in consideration. Note thatall the three NE types (PERSON, ORGANIZATION, LOCATION) in the Name category are among the fivemost frequent types. In the Time category, DATE type is much more significant than TIME and PERIODtypes, probably because the significance of a piece of news is characterized by the scale of a day, a monthor a year, rather than a specific time of a day, or a period of time. In the Number category, the general Num-ber type (NUMBER) of named entities is overwhelmingly more significant than all the other specific Numbertypes (such as MONEY, POWER, LENGTH, etc.).

Among these five types of NEs, three (PERSON, LOCATION and DATE) of them are more importantthan the other two (NUMBER and ORGANIZATION) for separating relevant sentences from non-relevantsentences. For example, for TREC 2003, the percentage of relevant sentences that include PERSON typename entities, PR(PERSON), is about 13% more than that of the non-relevant sentences, PNR(PERSON).The discrimination capability of the ORGANIZATION type in relevance detection is not as significant;PR(ORGANIZATION) is just marginally higher than PNR(ORGANIZATION). The insignificant role ofORGANIZATION in relevant retrieval has also been validated by our experiments on real data. One of

se cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,rmation Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 12: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

REC

TED

PR

OO

F461

462

463

464

465

466

467

468

469

470

471

472

473

474

475

476

477

Table 5.4The statistics of named entities (2002, 2003)

TREC 2002 Novelty Track, Total S# = 57227, MR = 1365,MNR = 55862

TREC 2003 Novelty Track, Total S# = 39820, MR = 15557,MNR = 24263

NEs Rel # (%) Non-Rel # (%) NEs Rel # (%) Non-Rel # (%)

Name category

PERSON 381 (27.91%) 13101 (23.45%) PERSON 6633 (42.64%) 7211 (29.72%)ORGANIZATION 532 (38.97%) 17196 (30.78%) ORGANIZATION 6572 (42.24%) 9211 (37.96%)LOCATION 536 (39.27%) 11598 (20.76%) LOCATION 5052 (32.47%) 5168 (21.30%)

Time category

DATE 382 (27.99%) 6860 (12.28%) DATE 3926 (25.24%) 4236 (17.46%)TIME 9 (0.66%) 495 (0.89%) TIME 140 (0.90%) 1154 (4.76%)PERIOD 113 (8.28%) 2518 (4.51%) PERIOD 1017 (6.54%) 705 (2.91%)

Number category

NUMBER 444 (32.53%) 14035 (25.12%) NUMBER 4141 (26.62%) 6573 (27.09%)ORDEREDNUM 77 (5.64%) 1433 (2.57%) ORDEREDNUM 725 (4.66%) 688 (2.84%)ENERGY 0 (0.00%) 5 (0.01%) ENERGY 0 (0.00%) 0 (0.00%)MONEY 66 (4.84%) 1775 (3.18%) MONEY 451 (2.90%) 1769 (7.29%)MASS 31 (2.27%) 1455 (2.60%) MASS 34 (0.22%) 19 (0.08%)POWER 16 (1.17%) 105 (0.19%) POWER 0 (0.00%) 0 (0.00%)TEMPERATURE 3 (0.22%) 75 (0.13%) TEMPERATURE 25 (0.16%) 9 (0.04%)DISTANCE 8 (0.59%) 252 (0.45%) DISTANCE 212 (1.36%) 47 (0.19%)SPEED 0 (0.00%) 0 (0.00%) SPEED 32 (0.21%) 2 (0.01%)LENGTH 46 (3.37%) 682 (1.22%) LENGTH 103 (0.66%) 29 (0.12%)HEIGHT 2 (0.15%) 25 (0.04%) HEIGHT 1 (0.01%) 3 (0.01%)AREA 5 (0.37%) 72 (0.13%) AREA 17 (0.11%) 11 (0.05%)SPACE 2 (0.15%) 54 (0.10%) SPACE 11 (0.07%) 10 (0.04%)PERCENT 62 (4.54%) 1271 (2.28%) PERCENT 371 (2.38%) 1907 (7.86%)Object CategoryURL 0 (0.00%) 0 (0.00%) URL 0 (0.00%) 62 (0.26%)

Others category

No NEs 246 (18.02%) 15899 (28.46%) No NEs 3272 (21.03%) 5533 (22.80)No POLD 359 (26.3%) 22689 (40.62%) No POLD 4333 (27.85%) 8674 (35.75%)No PLD 499 (36.56%) 31308 (56.05%) No PLD 6035 (38.79%) 12386 (51.05%)

12 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

UN

CO

Rthe reasons is that an NE of this type often indicates the name of a news agency; and names of news agenciesoften occur frequently in news articles. The role of the NUMBER type is not consistent between the twoTREC datasets (2002 and 2003). In Table 5.4, PR(NUMBER) is higher than PNR(NUMBER) for TREC2002, but is lower for TREC 2003. Therefore only the three effective types will be incorporated into the sen-tence retrieval step to improve the performance of relevance. This leads to our first observation on named enti-ties (NEs):

NE Observation #1. Named entities of the PLD types – PERSON, LOCATION and DATE are more effective

in separating relevant sentences from non-relevant sentences.

However, the ORGANIZATION type will be also used in the new pattern detection step since a differentorganization may provide new information. Here is an example

Example 3. Organization in new pattern detection

Ædocid = ‘‘NYT19980629.0465’’ num = ‘‘42’’æÆENAMEX TYPE = ‘‘ORGANIZATION’’æ Christian Coalition of FloridaÆ/ENAMEXæ director ÆENA-MEX TYPE = ‘‘PERSON’’æJohn DowlessÆ/ENAMEXæ disagrees with ÆENAMEX TYPE = ‘‘PER-SON’’æCarresÆ/ENAMEXæ’ spin but agrees that a judge or judges somewhere, state or federal, is liableto strike down the law.

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 13: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

478

479

480

481

482

483

484

485

486

487

488

489

490

491

492

493

494

495

496

497

498

499

500

501

502

503

504

505

506

507

508

509

510

511

512

514514

515

516

517

518

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 13

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

RO

OF

In this example, the organization name ‘‘Christian Coalition of Florida’’ gives an indication that new infor-mation may be provided by this sentence. We summarize this into NE Observation #2:

NE Observation #2: Named entities of the POLD types – PERSON, ORGANIZATION, LOCATION, and

DATE will be used in new pattern detection; named entities of the ORGANIZATION type may provide dif-ferent sources of new information.

Table 5.4 also lists the statistics of those sentences with no NEs at all, with no POLD (Person, Organiza-tion, Location or Date) NEs, and with no PLD (Person, Location or Date) NEs. These data show the follow-ing facts: (1) There are obvious larger differences between relevant and non-relevant sentences without PLDNEs, than the differences without POLD NEs, or without any NEs. This confirms that PLD NEs are moreeffective in re-ranking the relevance score (NE Observation #1). (2) There are quite large percentages of rel-evant sentences without NEs or without POLD NEs. Therefore, first, the absence of NEs is not going to beused exclusively to remove sentences from the relevant sentence list. Second, the number of the previouslyunseen POLD NEs only contributes partly to novelty ranking.

P

OR

REC

TED

5.4. Named entities in event and opinion topics

As we have discussed, topics in TREC 2003 and TREC 2004 novelty track data collections are classifiedinto two types: opinion topics and event topics. If a topic can be transformed into multiple NE-questions,no matter it is an opinion or event topic, the relevant and novel sentences for this ‘‘specific’’ topic can beextracted by mostly examining required named entities (NEs) as answers to these questions generated fromthe topic. Otherwise, we can only treat it as a general topic for which no specific NEs can be used to identifysentences as answers. The analysis in Section 5.2 shows that we can use opinion patterns to identify opinionsentences that are more probably relevant to opinion topics (queries) than non-opinion sentences. However,opinion topics only consist of part of the queries (Table 5.5). There are only 22 opinion topics out of the 50topics from the 2003 novelty track. The number is 25 out of 50 for the 2004 novelty track. Now the question is:how to deal with those event topics?

Since an event topic (query) is usually about an event with persons, locations and dates involved, naturallywe turn to named entities about the types of PERSON, LOCATION and DATES for help. Statistics also ver-ify this hypothesis. Table 5.6 compares the difference in the statistics of NEs between event topics and opiniontopics for the TREC 2003 novelty track. In Table 5.6, each item lists two measurements: NY-Z(X), and PY-

Z(X). The former is the number of Y-type (Y is either R for relevant or NR for non-relevant) sentences forZ-category topics (Z is either O for opinion, or E for event), each of which has at least one of X-type namedentity (X is PERSON, LOCATION or DATE). The latter is the percentage of those sentences among all therelevant (or non-relevant) sentences in each category (event or opinion). These two measurements are calcu-lated as

TableKey da

R: Rel

OpinioEvent

PleaInfo

CNY�ZðX Þ ¼ # of Y -type sentences for Z-category topics with at least one X � type NE ð5:3ÞP Y�ZðX Þ ¼ NY�ZðX Þ=MY�Z ð5:4Þ

UNIn Eq. (5.4), MY-Z denotes the total number of Y-type (Y is either R for relevant or NR for non-relevant) sen-

tences in the Z category (Z is either E for event or O for opinion). Those numbers are listed in Table 5.5. ForPERSON, LOCATION and DATE types, the relevant to non-relevant percentage ratio for event topics, i.e.,PR_E(X)/PNR_E(X), is close to 2:1 on average, whereas the ratio for opinion topics, i.e., PR_O(X)/PNR_O(X), is

5.5ta about opinion and event topics (TREC 2003)

evant; NR: Non-Relevant # Topics # Sentences MR-Z MNR-Z

n topics (Z = O) 22 21,115 7755 13,360topics (Z = E) 28 18,705 7802 10,903

se cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,rmation Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 14: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

OF

519

520

521

522

523

524

525

526

527

528

529

530

531

532

533

534

Table 5.6Statistics of named entities in opinion and event topics (2003)

TREC 2003 Novelty Track Event Topics, Total = 18705,Rel# = 7802, Non-Rel# = 10,903

TREC 2003 Novelty Track Opinion Topics, Total S# = 21,115,Rel# = 7755, Non-Rel# = 13,360

NEs Rel # (%) Non-Rel # (%) NEs Rel # (%) Non-Rel # (%)

PERSON 3833 (49.13%) 3228 (29.61%) PERSON 2800 (36.11%) 3983 (29.81%)LOCATION 3100 (39.73%) 2567 (23.54%) LOCATION 1952 (25.17%) 2601 (19.47%)DATE 2342 (30.02%) 1980 (18.16%) DATE 1584 (20.43%) 2256 (16.89%)

14 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

PR

Oonly about 5:4 on average. Therefore, we obtain the following observation about named entities and opinion/event topics:

NE/OP Observation #3: Named Entities of the PERSON, LOCATION and DATE types play a more impor-tant role in event topics than in opinion topics.

This is further verified in our experiments of relevance retrieval. In the equation for NE-adjustment (Eq.(6.5)), the best results are achieved when a takes the value of 0.5 for event topics and 0.4 for opinion topics.Note that while opinion patterns in sentences are used for improving the performance of novelty detection foropinion topics, named entities play a more important role for event topics.

D

EC

TE6. Pattern-based approach to novelty detection

The unified information-pattern-Based Approach to Novelty Detection (ip-BAND) is illustrated in Fig. 1.There are three important steps in the proposed approach: query analysis, relevant sentence retrieval and novelsentence extraction. In the first step, an information request from a user will be implicitly transformed into oneor more potential questions in order to determine corresponding query-related information patterns. Informa-tion patterns are represented by combinations of query words and required answer types to the query. In thesecond step, sentences with the query-related patterns are retrieved as answer sentences. Then in the third step,

UN

CO

RR

Step1. Query Analysis (Sec. 6.1) (1) Question Generation

a. One or more specific questions, or b. A general question

(2) Information Pattern (IP) Determination a. For each specific question, IP entities includes

(a specific NE type, sentence length (SL) ) b. For a general question, IP entities includes

(POLD NE types, SL, and opinion/event type)

Step2. Relevant Sentence Retrieval (Sec. 6.2) - Sentence Relevance Ranking using TFISF Scores - Information Pattern Detection in Sentences - Sentence Re-Ranking using Information Patterns

Step3. Novel Sentence Extraction (Sec. 6.3) - For specific topic: New and Specific NE Detection - For general topic: New NEs and New Words Detection

Query (Topic)

Information Patterns

Relevant Sentences

Novel Sentences

Fig. 1. ip-BAND: a unified information-pattern-based approach to novelty detection.

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 15: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

535

536

537

538

539

540

541

542

543

544

545

546

547

548

549

550

551

552

553

554

555

556

557

558

559

560

561

562

563

564

565

566

567

568

569

570

571

572

573

574

575

576

577

578

579

580

581

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 15

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

sentences that indicate potential new answers to the questions are identified as novel sentences. We will detailthe three steps in Sections 6.1–6.3.

NC

OR

REC

TED

PR

OO

F6.1. Query analysis and question formulation

6.1.1. Query analysis

In the first step, a question formulation algorithm first tries to automatically formulate multiple specificNE-questions for a query, if possible. Each potential question is represented by a query-related pattern, whichis a combination of a few query words and the expected answer type. A specific question would require a par-

ticular type of named entities (NEs) for answers. Five types of specific questions are considered in the currentsystem: PERSON, ORGANIZATION, LOCATION, NUMBER and DATE.

If this is not successful, a general question will be generated. A general question does not require a particulartype of named entity for an answer. Any type of named entities (NEs) as listed in Table 5.4 as well as othersingle words could be a potential answer or part of an answer, as long as the answer context is related to thequestion. This means answers for general questions could also be in sentences without any named entities(NEs). However, from our data analysis, the NEs of POLD types (PERSON, ORGANIZATION, LOCA-

TION, DATE) are the most effective in detecting novel sentences, and three of them (PERSON, LOCATION,

DATE) are the most significant in separating relevant sentences from non-relevant sentences. In addition, aswe have observed in the statistics in Section 4, sentence lengths and opinion patterns are also important inrelevant sentence retrieval and novel sentence extraction. In particular, we can use opinion patterns to identifyopinion sentences that are more probably relevant to opinion topics (queries) than non-opinion sentences.PERSON, LOCATION and DATE play a more important role in event topics than in opinion topics. There-fore, for a general question, its information pattern includes topic type (event or opinion), sentence length,POLD NE types, as well as query words.

6.1.2. Question formulation

There are 49 queries in the TREC 2002 novelty track, 50 queries in the TREC 2003 novelty track and 50queries in the TREC 2004 novelty track. Our question formulation algorithm formulates multiple specificquestions for 8 queries from the TREC 2002 novelty track, for 15 queries from the TREC 2003 novelty track,and for 11 queries from the TREC 2004 novelty track, respectively. The remaining queries were transformedinto general questions.

Each query from the TREC novelty tracks has three fields: title, description and narrative. Even though notexplicitly provided in the format of questions, a significant number of the queries can be transformed into mul-tiple specific questions. As examples, query 306 from TREC 2002 novelty track and query N37 from TREC2003 are listed here as Examples 4 and 5.

Example 4. Question formulation – specific questions

ÆNumberæ: 306ÆTitleæ African Civilian DeathsÆDescriptionæ: How many civilian non-combatants have been killed in the various civil wars in Africa?ÆNarrativeæ: A relevant document will contain specific casualty information for a given area, country, orregion. It will cite numbers of civilian deaths caused directly or indirectly by armed conflict.

UExample 5. Question formulation – a general question

Æ Numberæ: N37ÆTitleæ Olympic scandal Salt Lake CityÆTopic Typeæ eventÆDescriptionæ: The Salt Lake City Olympic bribery scandal: What were the results?ÆNarrativeæ: Any mention of effects of the bribery scandal was relevant. Mention of things that should nothave been affected, such as, souvenir sales or attendance are relevant. Mention of members being involved

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 16: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

582

583

584

585

586

587

588

589

590

591

592

593

594

595

596

597

598

599

600

601

602

603 Q4

604

605

606

607

608

609

610

611

612

613

614

615

616

617

618

619

620

621

622

623

624

625

626

627

628

629

630

16 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

CTE

DPR

OO

F

in the scandal was relevant. Any mention of kinds of bribery used was relevant. Effects on potential donorsor sponsors of the SLC games or future games were relevant.

There are many approaches that can be used for question formulation and pattern determination. In ourcurrent implementation, we used a simple algorithm based on word-pattern matching to formulate questionsand corresponding information patterns from queries. For each type of the five NE-questions, a number ofword patterns, which could be unigram, bi-gram (or further n-gram) word sequences, have been constructedfor question type identification. Some word patterns were extracted from the TREC 2002 novelty track queriesmanually and some patterns were selected from our question answering system ( Li, 2003). The patterns havebeen shown Table 4.1 (in Section 4), where, most of them are unigram, but some of them are bi-gram. For agiven query, the algorithm will go through the text in both the description and the narrative fields to identifyterms that matches some word patterns in the list. The query analysis component first tries to formulate atleast two specific questions for each query, if possible, because a single specific question probably only coversa small part of a query, therefore it is translated into a general question.

For the topic in Example 4, in the Description and the Narrative fields, three word patterns ‘‘area’’, ‘‘coun-try’’ and ‘‘region’’ match the word patterns in the answer type Location, and two word patterns ‘‘how many’’and ‘‘numbers’’ matches the word patterns in the answer type Number. Therefore, two specific questionsabout ‘‘Where’’ and ‘‘How many’’ are formulated, hence the topic is identified as a specific topic.

If a query only has terms that match with patterns belonging to one type of question, or it does not haveany matched terms at all, then a general question is generated for the query. For a general topic, the topic type(event or opinion), as part of the information pattern, is given in the TREC 2003 and 2004 novelty tracks. Anumber of expected opinion patterns have been identified manually and are listed in Table 4.2 (in Section 4).For the topic in Example 5, in the Narrative fields, a word pattern ‘‘member’’ matches the one in the answertype Person. Because only one specific question about ‘‘Who’’ is formulated, the topic is classified as a generaltopic.

Question formulation is more promising for long and well-specified queries. However, this approach is alsoapplicable to short queries with just a few key words. In the latter case, a query will most probably be trans-formed into a general question if sufficient information is not available from such a short query.

E

UN

CO

RR6.2. Using patterns in relevance re-ranking

6.2.1. Relevant sentence retrieval overviewThe task of relevant sentence retrieval is to retrieve sentences that are likely to have potential answers

to the questions transformed from a given topic. It first takes query words of the query, and searches inits data collection to retrieve sentences that are topically relevant to the query. Our relevant sentenceretrieval module is implemented based on an existing search toolkit –LEMUR (Lemur, 2006). The modulefirst uses a ‘‘TFISF’’ model adopted from TFIDF models in LEMUR for the first round of relevant sen-tence retrieval. Then it removes the sentences that do not contain any ‘‘answers’’ to the potential question.For a specific question, only a specific type of named entity that the question expects would be consideredfor potential answers. Thus a sentence without an expected type of named entity will be removed from thelist. A list of presumed answer sentences (which contain expected named entities to the question) is gen-erated. For general questions (topics), all types of named entities as well as single words could be potentialanswers. Therefore, no filtering is performed after the first round of sentence retrieval using the TFISFmodel.

To improve the performance of finding relevant sentences and to increase the ranks for sentences with therequired information patterns, sentence-level information patterns, including sentence lengths, Person-Loca-tion-Date NEs, and opinion patterns, are incorporated in the relevance retrieval step to re-rank the retrievedsentences. The use of information patterns seems to have attracted attention for other applications. Recentwork by Kumaran and Allan (2006) tried a pattern-based approach for document retrieval. By asking userssimple questions, frequent bi-grams within a window of eight terms and phrases in queries were identified andused to improve pseudo relevance feedback results.

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 17: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

631

632

633

634

635

636

637

638

639

640

642642

643

644

645

647647

648

649

650

651

653653

654

655

656

657

658 Q3

659

660

661

662

663

664

665

666

667668

670670

671

672

673

674

675

676677

679679

680

681

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 17

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

OF

6.2.2. Ranking with TFISF models

TFIDF models are one of the typical techniques in document retrieval (Salton & McGill, 1983). TF standsfor Term Frequency in a document and IDF stands for Inverse Document Frequency with respect to a doc-ument collection. We adopt TFIDF models for the relevant sentence retrieval step in our novelty detectiontask simply because it was also used in other systems and was reported to be able to achieve equivalent orbetter performance compared to other techniques in sentence retrieval (Allan et al., 2003). The name ofour sentence retrieval model is called the TFISF model, to indicate that inverse sentence frequency is usedfor sentence retrieval instead of inverse document frequency. The initial TFISF relevance ranking score S0

for a sentence, modified from the LEMUR toolkit (Lemur, 2006; Zhai, 2001), is calculated according tothe following formula:

PleaInfo

OS0 ¼Xn

i¼0

½wðtiÞtfsðtiÞtfqðtiÞisf 2ðtiÞ� ð6:1Þ

PRwhere n is the total number of terms, isf(ti) is inverse sentence frequency (instead of inverse document fre-

quency in document retrieval), tfs(ti) is the frequency of term ti in the sentence, and tfq(ti) is the frequencyof term ti in the query. The inverse sentence frequency is calculated as

isf ðtiÞ ¼ logNNti

ð6:2Þ

EDwhere N is the total number of sentences in the collection, N ti is the total number of sentences that include the

term ti. Note that in the above formulas, ti could be a term in the original query (with a weight w(ti) = 1) or inan expanded query that has more terms from pseudo feedback (with a weight w(ti) = 0.4). The expandedquery with pseudo relevance feedback is represented by

TfXWqg ¼ fWqg þ fTop 50 most frequent words from the top 100 sentences retrievedg ð6:3Þ

CO

RR

ECWith pseudo relevance feedback, the system assumes that top 100 sentences retrieved are relevant to the query

and top 50 most frequent terms within the 100 sentences are added to the original query. We have tried variousnumbers of top ranked sentences and top most frequent terms and have found that 50 top words in the 100 topsentences give the best performance. As in other information retrieval systems, stopword removal (e.g., Zhanet al., 2003; Li, 2003) and word stemming (Krovetz, 1993) are performed in a preprocessing step for relevantsentence retrieval.

The score S0 will be served as the baseline for comparing the performance increase with the informationpatterns we have proposed in relevant sentence retrieval and novelty detection. We have tried both the TFISFmodel with original queries and the TFISF model with expanded queries by pseudo feedback. The TFISFmodel with pseudo feedback is used in the experiments reported in the rest of the paper because it providesbetter performance.

6.2.3. TFISF with information patterns

The TFISF score is adjusted using the following three information patterns: sentence lengths, named enti-ties, and opinion patterns. The length-adjustment is calculated as

S1 ¼ S0 � ðL=LÞ ð6:4Þ

UNwhere L denotes the length of a sentence and L denotes the average sentence length. In the current implemen-

tation, the average sentence length is set as the same for all topics in each collection of TREC 2002, 2003 and2004 novelty tracks. This parameter is not critical as long as the comparison is only among sentences for thesame topic. A sentence that is longer than the average will increase the TFISF score. This adjustment followsSL Observations #1 and #2.

The NEs-adjustment is computed as

S2 ¼ S1 � ½1þ aðF person þ F location þ F dateÞ� ð6:5Þ

where Fperson = 1 if a sentence has at least a person’s name, 0 otherwise; Flocation = 1 if a sentence has at least alocation name, 0 otherwise; and Fdate = 1 if a sentence has at least a date, 0 otherwise. This adjustment follows

se cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,rmation Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 18: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

682

683

684

685686

688688

689

690

691

692

693

694

695

696

697

698

699

700

701

702

703

704

705

706

707

708

709

710

711

712

713

714

715

716

717

718

719

720

721

722

723

724

725

726

727

728

729

730

731

732

18 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

NE Observation #1. The adjustment is applied to both specific topics and general topics, but the parameter ais set slightly different. For specific topics or general but opinion topics, a = 0.4; for general, event topics,a = 0.5. This follows the NE/OP Observation #3.

Finally, the opinion-adjustment is computed as

PleaInfo

S3 ¼ S2 � ½1þ bF opinion� ð6:6Þ

OF

where Fopinion = 1 if a sentence is identified as an opinion sentence with one or more opinion patterns, 0 other-wise. A number of patterns (i.e., ‘‘said’’, ‘‘argue that’’, see Table 5.2) are used to determine whether a sentenceis an opinion sentence. This adjustment follows the OP Observations #1 and #2. This final adjustment stepbased on opinion patterns is only performed for general opinion topics.

O

UN

CO

RR

EC

TED

PR6.2.4. Formula/parameter optimization

All three adjustments are applied to the training data based on TREC 2002 and/or TREC 2003 noveltytrack to find the best set of parameters, a and b, and the same set of parameters are used for all data sets.Here is a brief summary of the formula/parameter optimization procedure; details can be found in Li (2006).

(1) First, we have tried different ways of adjusting scores and found that above adjustment mechanism (Eqs.(6.4)–(6.6)) achieves the best performance.

In incorporating the information of sentences lengths into relevant sentence re-ranking, we have tried thefollowing methods: +Length method (to ‘‘add’’ length factor to the original score), *ClippedLength method (to‘‘multiple’’ a roof-clipped length factor – penalty is given to sentences shorter than average) and *Length

method (to ‘‘multiple’’ a length factor). They are all compared with the baseline, in which the original TFISFranking score S0 is simply normalized by the maximum possible ranking score for all sentences in consider-ation. We carried out performance comparison with different sentence length adjustment methods to 49 topicsfrom TREC 2002, and it turns out that *Length method outperforms the other two significantly. Therefore weuse the *Length method in our ip-BAND approach.

The following six methods are compared with the baseline with normalized relevant ranking S0: +ALL (to‘‘add’’ the occurrence of different types of all the NEs), +POLD (to ‘‘add’’ occurrence of the POLD NEs),+PLD (to ‘‘add’’ occurrence of the PLD NEs), *ALL(to ‘‘multiple’’ the occurrence of different types of allthe NEs), *POLD(to ‘‘multiple’’ occurrence of the POLD NEs), *PLD (to ‘‘multiple’’ occurrence of thePLD NEs). Based on the performance comparison with different named entities adjustment methods to 49topics from TREC 2002, the *PLD method outperforms the all the other five methods significantly. Note thatwe have not only compared different ways of applying named entities, i.e., ‘‘adding’’ or ‘‘multiplying’’, but alsotried different combinations of named entities. It is clear that using only Person/Location/Data-type namedentities gives the best performance, with the two different calculations. Between the two ways of calculation,‘‘multiplying’’ method is clearly better. Therefore we use the *PLD method in our ip-BAND approach.

In incorporating the information of opinion patterns into relevant sentence re-ranking, we have tried thefollowing methods: +Opinion method (to ‘‘add’’ a factor if opinion patterns exist) and *Opinion method(to‘‘multiple’’ a factor if opinion patterns exist). They are both compared with the baseline with original TFISFranking score S0 normalized by the maximum possible ranking score for all sentences in consideration. Basedon the performance comparison with the two different opinion pattern adjustment methods to 23 topics fromTREC 2003, the precisions in relevant sentence retrieval by using the two methods are very close. And we havealso found that different values of this parameter do not change the performance very much. Therefore eitherone of them could be chosen into our ip-BAND approach. However, to be consistent with the other twoadjustments (the length adjustment and the NE-adjustment), we select the *Opinion method in our currentimplementation.

(2) After selecting the most appropriate formulas in the above studies, we applied all the three adjustmentsand tune the parameters for the best performance of relevant sentence retrieval based on the TREC 2003dataset.

To simplify the computation, we apply the three adjustments sequentially. The sentence length adjustmentis applied first using Eq. (6.4). No parameters need to be tuned for this adjustment. Then the NE-adjustmentis applied with different values of the parameter a, ranging from 0.1 to 1.0, with an interval of 0.1, using

se cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,rmation Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 19: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

733

734

735

736

737

738

739

740

741

742

743

744

745

746

747

748

749

750

751

752

753

754

755

756

757

758

759

760

761

762763765765

766

767

768

769

770

771

772

773

774

775

776

777

778

779

780

781

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 19

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

REC

TED

PR

OO

F

Eq. (6.5). It turns out that for general, opinion topics, the best performance is achieved when a = 0.5, and forgeneral, event topics, or specific topics, a = 0.4 gives the best results.

Finally the opinion pattern adjustment is applied with different values of b, ranging from 0.1 to 1.0, with aninterval of 0.1, using Eq. (6.6). We have found that for specific topics, we could tune a parameter for the bestperformance of relevant sentence retrieval. However, the performance in the final novelty detection is betterwithout the opinion pattern adjustment. Therefore, the opinion pattern adjustment is not used for specifictopic. For general opinion topics, the best performance is achieved when b = 0.5.

Incorporating information patterns at the retrieval step should improve the performance of relevance andthus help in the following novelty extraction step. After applying the above three steps of adjustments on theoriginal ranking scores, sentences with query-related information patterns are pulled up in the ranked list. Forthe following two sentences in Example 2 in Section 1, the relevant (and novel) sentence (sentence 1) wasranked 14th with the original TFISF ranking scores. It was pulled up to the 9th place in the ranked list afterthe adjustments with the information patterns. The non-relevant sentence (sentence 2) was initially ranked2nd, but pushed down to the 81st place after the score adjustments. Complete comparison results of noveltydetection performance on TREC 2002, 2003 and 2004 are provided in the experiments in the next section.

Even though the parameters are trained on TREC 2002 and/or 2003 data, we have also tested the generalityof these parameters. We have found that the best parameters that trained from TREC 2002 and 2003 data arealso the best for the TREC 2004 data. This is verified by applying all the different sets of parameters used inthe training step to the TREC 2004 data. This indicates that training the parameters with TREC 2004 data willyield the same set of optimal parameters.

6.3. Novel sentence extraction

We start with our treatment for specific topics for easy explanation. For a specific topic with multiple spe-cific NE-questions that are formulated from the topic (query), an answer sentence may have answers to one ormore of these specific NE-questions. So named entities related to any one of the specific questions in theanswer sentences should be extracted, and are treated as answers. There is an answer pool associated with eachtopic, which is initially empty. As sentences come in, new answers (i.e., new and specific NEs) will be added tothe answer pool when the novel sentence detection module determines that the incoming answers are previ-ously unseen. A sentence will be marked novel if it contains new answers. Sentences without new answers willbe removed from the final list provided to the user.

For a general topic, the novelty score of a sentence is calculated with the following formula:

PleaInfo

Sn ¼ xNw þ cNne ð6:7Þ

UN

CO

Rwhere Sn is the overall novelty score of a sentence S, Nw is the number of new words (including NEs) in S thatdo not appear in its previous sentences, Nne is the number of POLD-type named entities in S that do not ap-pear in its previous sentences, and x and c are weights for new words and new named entities, respectively.Note that stopwords have been removed from all sentences in the relevant sentence step, and will not be con-sidered in the process of novelty score calculation. Then the words are stemmed. A sentence is identified as anovel sentence if its novelty score is equal to or greater than a preset threshold. In our experiments, for thegeneral topics, the best performance of novelty detection is achieved when both x and c are set to 1, puttingequal weights on the new words and new NEs, and the threshold T for Sn is set to 4, indicating at least four ofthem in total counts for a novel sentence.

The novelty score formula given in Eq. (6.7) is actually a general one that can also be applied to specifictopics. In that case, Nnes the number of the specific answer NEs, and we set x to 0 since only specific answerNEs are counted. The threshold for the novelty score Sn is set to 1, indicating the appearance of one specificNE makes the sentence novel.

7. Experimental results and analysis

In this section, we present and discuss the main experimental results. The data used in our experiments arefrom the TREC 2002, 2003 and 2004 novelty tracks. The comparison of our approach and several baseline

se cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,rmation Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 20: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

782

783

784

785

786

787

788

789

790

791

792

793

794

795

796

20 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

approaches are described. The parameters used in both the baselines and our approaches were tuned with theTREC 2002 data, except the parameter b in Eq. (6.6), which is used for sentence re-ranking for opinion topicsin our approach. The experiments and analysis include the performance of novelty detection for specific top-ics, general topics, and the overall performance for all topics. Issues of baseline selection and evaluation mea-sures are also discussed.

PR

OO

F

7.1. Baseline approaches

We compared our information-pattern-based approach to novelty detection (ip-BAND) to four baselinesdescribed in the following subsections. They are: B-NN: baseline with initial retrieval ranking (without noveltydetection), B-NW: baseline with new word detection, B-NWT: baseline with new word detection with a thresh-old, and B-MMR: baseline with maximal marginal relevance (MMR). For comparison, in our experiments,the same retrieval system based on the TFISF techniques adopted from the LEMUR toolkit (Lemur, 2006)is used to obtain the retrieval results of relevant sentences in both the baselines and our ip-BAND approach.The evaluation measure used for performance comparison is precision at rank N (N = 5, 10, 15, 20 and 30 inTables 7.1–7.12). It shows the fraction of correct novel (or relevant) sentences in the top N sentences deliveredto a user, which is defined as

UN

CO

RR

EC

TEDTable 7.1

Performance of novelty detection for 8 specific topics (queries) from TREC 2002

Top # sentences B-NN B-NW B-NWT B-MMR ip-BAND

Precision Precision Chg% Precision Chg% Precision Chg% Precision Chg%

5 0.1750 0.2000 14.3 0.1750 0.0 0.1750 0.0 0.2000 14.310 0.1625 0.1750 7.7 0.2000 23.1 0.1375 �15.4 0.2500* 53.8*

15 0.1667 0.1667 0.0 0.1833 10.0 0.1500 �10.0 0.2250* 35.0*

20 0.1750 0.1750 0.0 0.1875 7.1 0.1562 �10.7 0.1938 10.730 0.1625 0.1708 5.1 0.1792 10.3 0.1458 �10.3 0.1750 7.7

Note: Data with * pass significance test at 95% confidence level by the Wilcoxon test and ** for significance test at 90% level – same appliesto Tables 7.1–7.13.

Table 7.2Performance of novelty detection for 15 specific topics (queries) from TREC 2003

Top # sentences B-NN B-NW B-NWT B-MMR ip-BAND

Precision Precision Chg% Precision Chg% Precision Chg% Precision Chg%

5 0.5067 0.5867** 15.8** 0.5867** 15.8** 0.5200 2.6 0.6933* 36.8*

10 0.5400 0.5867 8.6 0.6467* 19.8* 0.5600 3.7 0.6933* 28.4*

15 0.4978 0.6133* 23.2* 0.6578* 32.1* 0.6000* 20.5* 0.6756* 35.7*

20 0.5200 0.6300* 21.2* 0.6667* 28.2* 0.5867* 12.8* 0.6800* 30.8*

30 0.5133 0.6022* 17.3* 0.6844* 33.3* 0.5978* 16.5* 0.7000* 36.4*

Table 7.3Performance of novelty detection for 11 specific topics (queries) from TREC 2004

Top # sentences B-NN B-NW B-NWT B-MMR ip-BAND

Precision Precision Chg% Precision Chg% Precision Chg% Precision Chg%

5 0.2000 0.2364 18.2 0.2364 18.2 0.2182 9.1 0.2727 36.410 0.2455 0.2364 �3.7 0.2455 0.0 0.2455 0.0 0.2909 18.515 0.2303 0.2545 10.5 0.2545 10.5 0.2364 2.6 0.3030* 31.6*

20 0.2455 0.2545 3.7 0.2773 13.0 0.2500 1.9 0.3182* 29.6*

30 0.2394 0.2515** 5.1** 0.2818 17.7 0.2727 13.9 0.3152* 31.6*

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 21: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

UN

CO

RR

EC

TED

PR

OO

F

Table 7.4Performance of novelty detection for 41 general topics (queries) from TREC 2002

Top # sentences B-NN B-NW B-NWT B-MMR ip-BAND

Precision Precision Chg% Precision Chg% Precision Chg% Precision Chg%

5 0.1902 0.1951 2.6 0.2049 7.7 0.2293 20.5 0.2390** 25.6**

10 0.2000 0.1951 �2.4 0.2049 2.4 0.2098 4.9 0.2098 4.915 0.1935 0.2000 3.4 0.2016 4.2 0.2033 5.0 0.2114 9.220 0.1854 0.1890 2.0 0.1939 4.6 0.1817 �2.0 0.2073 11.830 0.1748 0.1772 1.4 0.1707 �2.3 0.1691 �3.3 0.1902** 8.8**

Table 7.5Performance of novelty detection for 35 general topics (queries) from TREC 2003

Top # sentences B-NN B-NW B-NWT B-MMR ip-BAND

Precision Precision Chg% Precision Chg% Precision Chg% Precision Chg%

5 0.4229 0.4171 �1.4 0.4457 5.4 0.4343 2.7 0.5257* 24.3*

10 0.4143 0.4371** 5.5** 0.4657** 12.4** 0.4571 10.3 0.5257* 26.9*

15 0.4152 0.4400* 6.0* 0.4552** 9.6** 0.4438 6.9 0.5124* 23.4*

20 0.4057 0.4343* 7.0* 0.4686* 15.5* 0.4200 3.5 0.5029* 23.9*

30 0.3867 0.4238* 9.6* 0.4590* 18.7* 0.4219 9.1 0.4867* 25.9*

Table 7.6Performance of novelty detection for 39 general topics (queries) from TREC 2004

Top # Sentences B-NN B-NW B-NWT B-MMR ip-BAND

Precision Precision Chg% Precision Chg% Precision Chg% Precision Chg%

5 0.2359 0.2359 0.0 0.2410 2.2 0.2359 0.0 0.2154 -8.710 0.2026 0.2026 �0.0 0.2077 2.5 0.2026 0.0 0.2256 11.415 0.1949 0.2000 2.6 0.2051 5.3 0.1949 �0.0 0.2239* 14.9*

20 0.1859 0.1962* 5.5* 0.1974 6.2 0.1846 -0.7 0.2128* 14.5*

30 0.1735 0.1821** 4.9** 0.1846** 6.4** 0.1684 -3.0 0.1991* 14.8*

Table 7.7Performance of novelty detection for 49 queries (all topics) from TREC 2002

Top # sentences B-NN B-NW B-NWT B-MMR ip-BAND

Precision Precision Chg% Precision Chg% Precision Chg% Precision Chg%

5 0.1878 0.1959 4.3 0.2000 6.50 0.2204 17.4 0.2327** 23.9**

10 0.1939 0.1918 �1.1 0.2041 5.30 0.1980 2.1 0.2163 11.615 0.1891 0.1946 2.9 0.1986 5.00 0.1946 2.9 0.2136 12.920 0.1837 0.1867 1.7 0.1929 5.00 0.1776 �3.3 0.2051** 11.7**

30 0.1728 0.1762 2.0 0.1721 �0.40 0.1653 �4.3 0.1878** 8.7**

Table 7.8Performance of novelty detection for 50 queries (all topics) from TREC 2003

Top # sentences B-NN B-NW B-NWT B-MMR ip-BAND

Precision Precision Chg% Precision Chg% Precision Chg% Precision Chg%

5 0.4480 0.4680 4.5 0.4880 8.9 0.4600 2.7 0.5760 * 28.6*

10 0.4520 0.4820* 6.6* 0.5200* 15.0* 0.4880 8.0 0.5760* 27.4*

15 0.4400 0.4920* 11.8* 0.5160* 17.3* 0.4907* 11.5* 0.5613* 27.6*

20 0.4400 0.4930* 12.0* 0.5280* 20.0* 0.4700* 6.8* 0.5560* 26.4*

30 0.4247 0.4773* 12.4* 0.5267* 24.0* 0.4747* 11.8* 0.5507* 29.7*

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 21

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 22: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

UN

CO

RR

EC

TED

PR

OO

F

798798

799

800

801

Table 7.9Performance of novelty detection for 50 queries (all topics) from TREC 2004

Top # sentences B-NN B-NW B-NWT B-MMR ip-BAND

Precision Precision Chg% Precision Chg% Precision Chg% Precision Chg%

5 0.2280 0.2360 3.5 0.2400 5.3 0.2320 1.8 0.2280 0.010 0.2120 0.2100 �0.9 0.2160 1.9 0.2120 0.0 0.2400** 13.2**

15 0.2027 0.2120 4.6 0.2160 6.6 0.2040 0.7 0.2413* 19.1*

20 0.1990 0.2090* 5.0* 0.2150 8.0 0.1990 0.0 0.2360* 18.6*

30 0.1880 0.1973* 5.0* 0.2060* 9.6* 0.1913 1.8 0.2247* 19.5*

Table 7.10Comparison among specific, general and all topics at top 15 ranks

Data > TREC 2002 TREC 2003 TREC 2004

Topic n/ Chg% Nvl# Rdd# NRl# Chg% Nvl# Rdd# NRl# Chg% Nvl# Rdd# NRl#

Specific 35.0* 3.38 0.12 11.50 35.7* 10.13 3.07 1.8 31.6* 4.55 5.42 5.03General 9.2 3.17 0.27 11.56 23.4* 7.68 2.66 4.66 14.9* 3.35 3.34 8.31All 12.9 3.20 0.25 11.55 27.6* 8.42 2.78 3.8 19.1* 3.62 3.74 7.64

Chg%: Improvement over the first baseline in percentage; Nvl#: Number of true novel sentences; Rdd#: Number of relevant butredundant sentences; NRl#: Number of non-relevant sentences.

Table 7.11Performance of relevance for specific topics (queries) (a = 0.4 in Eq. (6.4))

TREC 2002 (8 topics) TREC 2003 (15 topics) TREC 2004 (11 topics)

Top # sentences TFISF Length + NEs TFISF Length + NEs TFISF Length + NEs

Precision Precision Chg% Precision Precision Chg% Precision Precision Chg%

5 0.1750 0.2000 14.3 0.8800 0.9333 6.1 0.6727 0.5636 -16.210 0.1750 0.2250 28.6 0.8733 0.9200 5.3 0.6545 0.6182 -5.615 0.1750 0.2000 14.3 0.8356 0.9067** 8.5** 0.6061 0.6485 7.020 0.1875 0.2062 10.0 0.8567 0.8867 3.5 0.6136 0.6591 7.430 0.1708 0.1833 7.3 0.8444 0.8756** 3.7** 0.6030 0.6182 2.5

Table 7.12Performance of relevance for general topics (queries)

TREC 2002 (41 topics) TREC 2003 (35 topics) TREC 2004 (39 topics)

Top # sentences TFISF Length + NEs(1) TFISF Length + NEs + Opinion(2) TFISF Length + NEs + Opinion(2)

Precision Precision Chg% Precision Precision Chg% Precision Precision Chg%

5 0.2049 0.2488** 21.4** 0.6629 0.7086 6.9 0.4615 0.4564 �1.110 0.2171 0.2220 2.2 0.6200 0.7000* 12.9* 0.4359 0.4615 5.915 0.2114 0.2260 6.9 0.6343 0.6857* 8.1* 0.4308 0.4462 3.620 0.2000 0.2159 7.9 0.6386 0.6714** 5.1** 0.4141 0.4410* 6.5*

30 0.1870 0.2033** 8.7** 0.6371 0.6552 2.8 0.4026 0.4342* 7.9*

Notes: (1) a = 0.4 (2) a = 0.5 for event topics, a = 0.4, b = 0.5 for opinion topics.

22 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

PleaInfo

precisionðNÞ ¼ number of novelðrelevantÞsentences retrieved

Nðnumber of sentence retrievedÞ ð7:1Þ

7.1.1. B-NN: initial retrieval ranking

The first baseline does not perform any novelty detection but only uses the initial sentence ranking scoresgenerated by the retrieval system directly as the novelty scores (without filtering and re-ranking). One purpose

se cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,rmation Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 23: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

802

803

804

805

806

807

808

809

810

811

812

813

814

815 Q5

816

817

818

819

820

821

822

823

824

825

826

827

829829

830

831

832

833

834

835

836

837

838

839

840

841

842

843

844

845

846

847

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 23

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

EC

TED

PR

OO

F

of using this baseline is to see how much novelty detection processes (including NE filtering, relevant sentencere-ranking and new answer detection) may help in removing redundancies.

7.1.2. B-NW: new word detection

The second baseline in our comparison is simply applying new word detection. Starting from the initialretrieval ranking, it keeps sentences with at least one new word that does not appear in previous sentencesas novel sentences, and removes those sentences without new words from the list. As a preprocessing step,all words in the collection were stemmed and stopwords were removed.

7.1.3. B-NWT: new word detection with a threshold

The third baseline (B-NWT) is similar to B-NW. The difference is that it counts the number of new wordsthat do not appear in previous sentences. A sentence is identified as novel sentence if and only if the number ofnew words is equal to or greater than a preset threshold. The best value of the threshold is 4 in our experi-ments. This means that a sentence is treated as a novel sentence only if more than three new words areincluded in the sentence. The threshold is selected as the same as for our ip-BAND novel sentence determina-tion for general topics (Eq. 6–11). So, roughly speaking, this baseline is comparable to the ip-BAND appliedto general topics.

7.1.4. B-MMR: Maximal Marginal Relevance (MMR)

Many approaches to novelty detection, such as maximal marginal relevance (MMR), new word count mea-sure, set difference measure, cosine distance measure, language model measures, etc., were reported in the lit-erature. The MMR approach was introduced by Carbonell and Goldstein (1998), which was used for reducingredundancy while maintaining query relevance in document re-ranking and text summarization. In our exper-iments, the MMR baseline approach (B-MMR) starts with the same initial sentence ranking used in otherbaselines and our ip-BAND approach. In B-MMR, the first sentence is always novel and ranked top in nov-elty ranking. All other sentences are selected according their MMR scores. One sentence is selected and putinto the ranking list of novelty sentences at a time. MMR scores are recalculated for all unselected sentencesonce a sentence is selected. The process stops until all sentences in the initial ranking list are selected. MMR iscalculated by

PleaInfo

MMR ¼ arg maxSi2R=N

½kðSim1ðSi;QÞ � ð1� kÞÞmaxSj2N

Sim2ðSi; SjÞ� ð7:2Þ

UN

CO

RR

where Si and Sj are the ith and jth sentences in the initial sentence ranking, respectively, Q represents thequery, N is the set of sentences that have been currently selected by MMR, and R/N is the set of sentenceshave not yet selected. Sim1 is the similarity metric between sentence and query used in sentence retrieval,and Sim2 can be the same as Sim1 or can be a different similarity metric between sentences.

We use MMR as a baseline because MMR was reported to work well in non-redundant text summarization(Carbonell & Goldstein, 1998), novelty detection at document filtering (Zhang, Callan et al., 2002) and sub-topic retrieval (Zhai et al., 2003). Also, MMR may incorporate various novelty measures by using differentsimilarity matrix between sentences and/or choosing different value of k. For instance, if the cosine similaritymetric is used for Sim2 and k is set to 0, then MMR would become the cosine distance measure reported inAllan et al. (2003).

7.2. Experimental results and comparisons

Our ip-BAND approach also uses the same initial relevance ranking as all the baselines. The novelty detec-tion performance of our ip-BAND is a combined impact of information patterns in sentence filtering for spe-cific topics, relevant sentence re-ranking and new pattern detection. However, in order to see what is theimpact of information patterns for the relevant sentence retrieval step and the novel sentence extraction step,we have also conducted experiments comparing relevant sentence retrieval. In the following subsection, wepresent the results of the comparisons of the performance of both relevance and novelty detection, and forboth specific and general topics. Finally we present the overall performance for all the topics in the three

se cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,rmation Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 24: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

848

849

850

851

852

853

854

855

856

857

858

859

860

861

862

863

864

865

866

867

868

869

870

871

872

873

874

875

876

877

878

879

880

881

882

883

884

24 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

novelty track data collections. From Tables 7.1–7.13, Chg% denotes the percent change of the precision of ourip-BAND approach (or the 2nd, 3rd or 4th baseline) compared to the first baseline. The Wilcoxon test is usedfor significance test, with * and ** indicating the 95% and 90% confidence levels, respectively.

CO

RR

EC

TED

PR

OO

F7.2.1. Experimental Results on Novelty Detection

A. Specific topics. First, we tested the information-pattern-based approach to novelty detection (ip-BAND)on specific topics. Tables 7.1–7.3 show comparisons for the top 5, 10, 15, 20 and 30 sentences. The precisionmeasurements in the tables are the averages of all the specific topics in each data collection at several differentranks. For example, the performance of our approach with specific questions beats the first baseline by morethan 30% at rank 15, which are 35.0%, 35.7% and 31.6%, for TREC 2002, 2003 and 2004, respectively. Incomparison, the best results combining all the three baselines (No 2, No 3 and No 4) are only 10.0% (fromN-NWT), 32.1% (from N-NWT) and 10.5% (from both N-NW and N-NWT), respectively. In another words,within the top 15 sentences, our approach obtains more novel sentences than all four baselines. The precisionvalues (see Tables 7.1–7.3) of our ip-BAND approach are 22.5%, 67.6% and 30.3%, respectively. Theseindicate that within the top 15 sentences, there are 3.4, 10.1 and 4.6 sentences on average that are correctlyidentified as novel sentences by the ip-BAND approach (from TREC 2002, 2003 and 2004 respectively).The numbers are 1.5, 4.8 and 1.6 for the combinations of the best results of baselines 2–4.

In comparison, New Word Detection with a Threshold (B-NWT) performs slightly better than New WordDetection (B-NW), but Maximal Marginal Relevance (B-MMR) does not. On the surface, for specific topics,the new word detection baseline approach and our ip-BAND approach use the similar strategy. That is, once anew word (new specific NE) appears in a sentence, it is declared as a novel sentence. However, a new specificNE in our ip-BAND approach may answer the right question, but a new word in B-NW will often not. There-fore B-NWT is better than B-NW in the sense that it needs more words (4 in our experiments) to declare asentence to be novel. The performance of the baseline based on Maximal Marginal Relevance in fact is worsethan the first baseline that simply uses the results from initial relevant ranking at low recalls. This could be dueto the fact that MMR uses a combined score to measure both relevance and redundancy, which are somewhatin conflict. In reducing redundancy, it might also decrease the performance in relevant ranking, which is veryimportant in novelty detection.

We have the following conclusion on the experimental results for specific topics:

Conclusion #1. The proposed approach outperforms all baselines for specific topics.

B. General topics. Tables 7.4–7.6 show the performance comparison of our ip-BAND approach with thefour baselines in novelty detection on those general topics that cannot be turned into specific NE-questions.

The precision values for the top 15 sentences with our ip-BAND approach for general questions of theTREC 2002, 2003 and 2004 data are 21.1%, 51.2% and 22.4%, respectively. This represents that on average,3.2, 7.5 and 3.4 sentences are correctly identified as novel sentences, among the top 15 recalls. Again, the pre-cision is the highest for the 2003 data since this track has highest ratio of relevant to non-relevant sentences.Compared to the first baseline, the performance increases are 9.2%, 23.4% and 14.9%, respectively and the

UNTable 7.13

Performance of relevance for all topics (queries)

TREC 2002 (49 topics) TREC 2003 (50 topics) TREC 2004 (50 Topics)

Top # sentences TFISF Length + NEs(1) TFISF Length + NEs + Opinion(2) TFISF Length + NEs + Opinion(2)

Precision Precision Chg% Precision Precision Chg% Precision Precision Chg%

5 0.2000 0.2408** 20.4** 0.7280 0.7760** 6.6** 0.5080 0.4800 -5.510 0.2102 0.2224 5.8 0.6960 0.7660* 10.1* 0.4840 0.4960 2.515 0.2054 0.2218 7.9 0.6947 0.7520* 8.3* 0.4693 0.4907** 4.5**

20 0.1980 0.2143 8.2 0.7040 0.7360* 4.5* 0.4580 0.4890* 6.8*

30 0.1844 0.2000* 8.5* 0.6993 0.7213** 3.1** 0.4467 0.4747* 6.3*

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 25: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

885

886

887

888

889

890

891

892

893

894

895

896

897

898

899

900

901

902

903

904

905

906

907

908

909

910

911

912

913

914

915

916

917

918

919

920

921

922

923

924

925

926

927

928

929

930

931

932

933

934

935

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 25

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

TED

PR

OO

F

improvements are significant. In contrast, the best improvements combing all the three baselines (2–4) are only5.0% (from B-MMR), 9.6% (from B-NWT) and 5.4% (from B-NWT), respectively, which are much lower thanthe results of our ip-BAND approach. In comparison, New Word Detection with a Threshold achievesbetter performance than New Word Detection for general topics. This is within our expectation becauseNew Word Detection is a special case of New Word Detection with a Threshold when the new word thresholdis set to 1. In addition, both B-NWT and ip-BAND use the same threshold of 4 ‘‘new words’’ to declare arelevant sentence to be novel, except that ip-BAND also add additional counts for POLD-type NEs. Further,Maximal Marginal Relevance is slightly better than New Word Detection and New Word Detection with a

Threshold on the 2002 data, but it is worse than New Word Detection with a Threshold on the 2003 and2004 data.

We can draw the following main conclusions from the results:

Conclusion #2. Our ip-BAND approach consistently outperforms all the baseline approaches across the three

data sets: the 2002, 2003 and 2004 novelty tracks, for general topics.

Conclusion #3. In comparison, the performance of our ip-BAND approach is slighter better for the specific

topics than the general topics.

The reason for Conclusion #3 is mainly due to the fact that specific, targeted questions and answers areextracted for specific topics. This also indicates that as a future work, more improvements can be expectedif more general topics can be turned into NE or other specific questions.

C. All topics. The overall performance comparisons of the unified pattern-based approach with the fourbaselines on all topics from the TREC 2002, 2003 and 2004 novelty tracks are shown in Tables 7.7–7.9, respec-tively. The most important conclusion is the following:

Conclusion #4. On average, the unified pattern-based approach outperforms all baselines at top ranks when all

topics are considered.

Significant improvements are seen with the 2003 topics. In the top 15 sentences delivered, our approachretrieves 8.42 (= 15 * 0.5613) novel sentences on average, while the four baseline approaches only retrieve6.60, 7.38, 7.74 and 7.36 novel sentences, respectively. As anticipated, the overall performance for all topics(including both specific and general ones – Tables 7.10–7.12) is slightly better than that for the general topics(Tables 7.7–7.9), since the precision for the specific topics (Tables 7.1–7.3) are slightly higher than the generalones.

This comparison is also summarized in Table 7.10 at top 15 ranks (sentences). In the table, the followingmeasures are listed for each case of specific, general and all topics, and for TREC 2002, 2003 and 2004: (1) theimprovements over the first baseline in percentage (Chg%); (2) the number of correctly identified novel sen-tences; (3) the number of relevant but redundant sentences; and (4) the number of non-relevant sentences.The last three numbers add up to 15. According to the results shown in this table, we have the followingimportant observations that could guide further improvements in both relevant and novel sentence detectionfor different data collections.

(1) There are 3.20 novel sentences, 0.25 redundant sentences and 11.55 non-relevant sentences within the top15 sentences retrieved for the TREC 2002 topics. It indicates that further improving the performance ofrelevant sentence retrieval is still a major task for the TREC 2002 topics.

(2) There are 8.42 novel sentences, 2.78 redundant sentences and 3.8 non-relevant sentences within the top15 sentences retrieved for the TREC 2003 topics. The performance for the TREC 2003 topics is muchbetter than the performance for the TREC 2002 topics; therefore there is less room for performanceimprovements on either relevant sentence retrieval or novel sentence extraction.

(3) There are 3.62 novel sentences, 3.74 redundant sentences and 7.64 non-relevant sentences within the top15 sentences retrieved for the TREC 2004 topics. The performance of identifying novel sentences for theTREC 2004 topics is only slightly better than that for the TREC 2002 topics. Further performance gaincan be obtained by improving relevant sentence retrieval and/or novel sentence extraction.

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 26: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

936

937

938

939

940

941

942

943

944

945

946

947

948

949

950

951

952

953

954

955

956

957

958

959

960

961

962

963

964

965

966

967

968

969

970

971

972

973

974

975

976

977

978

979

980

981

982

983

984

985

26 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

TED

PR

OO

F

7.2.2. Experimental analysis on relevant sentence retrieval

The second set of experiments is designed to investigate the performance gain of finding relevant sentenceswith the sentence re-ranking step. For the specific topics, the relevant sentence retrieval module re-ranks thesentences by the revised scores that incorporate the information of sentence lengths and PLD-type named enti-ties (i.e., Person-, Location- and Date-type NEs). In addition, sentences without the required named entitiesare also removed (i.e., filtered) before re-ranking.

We compare the performance of finding relevant sentences with and without filtering and re-ranking. Thecomparison results are given in Table 7.11. From this Table, we can see that the retrieval improvements arenoticeable for the TREC 2002 novelty tracks, the improvements for TREC 2003 are moderate, and for TREC2004 improvements are seen for ranks 15, 20 and 30. One reason for the more significant improvements onTREC 2002 data collection could be that the impact of NEs and information patterns are more effective sincethe percentage of relevant sentences in this data collection is very low.

Nevertheless, the results in Tables 7.1–7.3 have shown that the pattern-based approach significantly outper-forms all the four baselines at top ranks for identifying novel sentences, for all three novelty data collections.This indicates that our pattern-based approach makes a larger difference at the step of detecting novel sen-tences than at the step of finding relevant sentences for those specific topics, particularly from the TREC2003 and 2004 novelty tracks. The reason could be that TREC 2003 and 2004 novelty track collections exhib-ited greater redundancy than the TREC 2002 and thus has less novel sentences, therefore our information pat-terns make a big difference here. In combination, information patterns play a balanced role in the two majorsteps: relevance retrieval (for TREC 2002 in particular) and novelty detection (for TREC 2003 and 2004 inparticular).

We also want to see how information patterns can improve the performance of novelty detection for thosegeneral topics that cannot be easily turned into multiple specific NE-questions. As described in Section 6, allthe three types of information patterns are incorporated in the relevance retrieval step of novelty detection forgeneral topics. Table 7.12 gives the performance comparison of relevance retrieval with the original TFISFranking and with the adjustments using these sentence level information patterns for the TREC 2002, 2003and 2004 data, respectively.

The main conclusion here is that incorporating information patterns and sentence level features into TFISFtechniques can achieve much better performance than using TFISF alone. Significant improvements areobtained for the 2003 topics and the 2004 topics. For example, at rank 15, the incorporation of the informa-tion patterns increases the precision of relevance retrieval by 6.9%, 8.1% and 3.6%, for TREC 2002, 2003 and2004, respectively. This lays a solid ground for the next step – new information detection, and therefore forimproving the performance of novelty detection for those general topics.

The overall performance of relevance for all topics is given in Table 7.13. This table shows that re-rankingsentences with information patterns achieves better performance than using TFISF alone, for all topics. Read-ers can also compare the improvements between the specific and the general topics by comparing the results inTables 7.11 and 7.12. Generally speaking, information patterns play a greater role in relevance retrieval forgeneral topics than for specific topics. We believe this is mostly due to the incorporation of opinion patternsinto relevant sentence re-ranking for the general and opinion topics (for TREC 2003 and 2004).

We have also conducted experiments to apply opinion patterns to the relevance re-ranking for specific top-ics, and we have found that the performance is worse than the results without opinion patterns as shown inTable 7.4. There could be two reasons. (1) Answers for specific topics are mostly specified by the specificrequired named entities, no matter the sentences include opinion patterns or not. (2). Answers for general,opinion topics are often indicated by opinion patterns, as have been shown in the statistics of opinion patternsin Section 5 (OP Observations #1 and #2).

7.3. Baselines and evaluation revisited

The experimental results show that the proposed pattern-based approach (ip-BAND) gives the best perfor-mance among all approaches compared. But there is still much room left for further improvements. There area few factors that may affect the performance of our system. These include: (1) baseline selection and evalu-ation measure by human assessors; (2) misjudgment of relevance and/or novelty by human assessors and

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 27: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

986

987

988

989

990

991

992

993

994

995

996

997

998

999

1000

1001

1002

1003

1004

1005

1006

1007

1008

1009

1010

1011

1012

1013

1014

1015

1016

1017

1018

1019

1020

1021

1022

1023

1024

1025

1026

1027

1028

1029

1030

1031

1032

1033

1034

1035

1036

1037

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 27

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

TED

PR

OO

F

disagreement of judgments between the human assessors; (3) limitation and accuracy of question formula-tions. Discussions of these factors can be found in Li (2006). Here we want to discuss the issue of baselineselection and evaluation measure; future work on question formulations will be presented in the next section.

Relevance or novelty judgment files by human assessors are usually used as the basis for evaluating the per-formance of different systems. In the TREC novelty tracks, the judgment of novel sentences was based on aparticular set of relevant sentences in a presumed order. Unlike relevance measures, a novelty or redundancymeasure is asymmetric. The novelty or redundancy of a sentence Si depends on the order of sentences(S1, . . . ,Si�1) that the user has seen before this one. For the TREC novelty track data, only the judgmentsfor a particular set of sentences in a presumed order are available. To collect novelty judgments of each sen-tence with respect to all possible subsets of orders, a human assessor has to read up to 2N�1 subsets. It isimpossible to collect complete novelty judgments in reality.

There are two potential problems with the current judgment/evaluation file. First, it is not very accurate toevaluate a system’s performance if the ranked sentences of a novelty detection system have a different orderfrom the particular set. Second, a relevant sentence, pre-marked redundant by human assessors, could be trea-ted as a novel sentence if it has new information that is not covered by previous sentences in the list providedby a novelty detection system. That is the case where the sentences that make it redundant in the list by humanassessors are simply not retrieved by the system.

One possible solution to this problem is to reorder the retrieved sentences following the natural order givenby the task. Note that this re-organization of the retrieved sentences was also done by NIST before presentingthe sentences to the assessors for novelty assessments of the TREC novelty tracks. It seems that the re-orga-nization of the retrieved sentences is a more appropriate (and stronger) baseline than the B-NN we have used.Further, the input to the novelty detection step of any novelty detection method should be re-organized in thesame way. Actually, this re-organization was also done by Allan et al. (2003) before testing the novelty meth-ods (including new word detection method). In our initial experiments, we have also tried the re-organizationidea. However, by doing that, the baseline approach without any novelty detection is even better than somenovelty methods (e.g., the new word detection method). At a first look, it seems that the problem came fromthe novelty detection step, but in fact it was mainly due to the low retrieval rate in the relevance retrieval step.The novelty detection step is to find new information from sentences that are supposed to be relevant. How-ever, if many sentences retrieved in the first step are not relevant, the novelty detection precision will be lowsince some non-relevant sentences may be treated as novel. Re-ordering of the retrieval sentences make thesituation worse since the relevance ranking of the reordered sentences does not make sense any more, andthe results in novelty detection are really not predictable.

An alternative approach is to deal with the problematic evaluation measures by human assessors. Somework has been done in this respect. For novelty detection data collected and used in their studies, researchersat CMU (Zhang et al., 2002) initially intended to collect judgments for 50 topics, but unfortunately they couldonly get assessments for 33 topics. They provide the information on which documents before a documentmakes it redundant. The documents must be listed in chronological order. Thus there are problems when eval-uating a novelty detection system in which documents are not output in chronological order. As research inter-est increases in novelty detection, more accurate and efficient data collection is crucial to the success ofdeveloping new techniques in this area.

One possible solution to a more appropriate novel sentence judgment is to generate dynamic novelty judg-ment files at evaluation time. Given a topic, instead of creating a fixed set of novel sentences based on the set ofrelevant sentences in a presumed order, relevant sentences can be classified into novelty groups. In each group,all sentences basically contain the same information with respect to the topic. Therefore, if a user reads anysentence from a novelty group, the rest of the sentences in the group will become redundant sentences to theuser because they do not contain any new information.

To evaluate the performance of a novelty detection system, a dynamic novelty judgment file can be gener-ated at evaluation time. Given the ranked list of sentences from the system and the set of relevant sentencesclassified in novelty groups for a topic, the dynamic novelty judgment file can be generated by simply scanningthe sentences in the ranked list and selecting sentences that are in the novelty groups of the topic and are thefirst sentence appearing in the ranked list from each novelty group. With the dynamic judgment file, the stan-dard evaluation tools provided by TREC can still be used to evaluate the performance of the novelty detection

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 28: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

1038

1039

1040

1041

1042

1043

1044

1045

1046

1047

1048

1049

1050

1051

1052

1053

1054

1055

1056

1057

1058

1059

1060

1061

1062

1063

1064

1065

1066

1067

1068

1069

1070

1071

1072

1073

1074

1075

1076

1077

1078

1079

1080

1081

1082

1083

1084

1085 Q1

28 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

TED

PR

OO

F

system. Evaluations with the dynamic novelty judgment file should be more accurate than using a fixed nov-elty judgment file based on a presumed order of sentences. An ideal system should select one sentence fromeach novelty group and deliver these sentences to the user. The ranked list of sentences provided by suchan ideal system will be a complete set of novel sentences for the topic in the sense that it contains all relevantinformation related to the topic. The performance of the ideal system is then 100% for both precision andrecall.

8. Conclusions and discussions

Novelty detection is an important task to reduce the amount of redundant as well as non-relevant informa-tion presented to a user. It is an important component of many potential applications, such as new eventdetection, document filtering, and cross-document summarization. In this paper, we introduce a new definitionof novelty: ‘‘Novelty or new information means new answers to the potential questions representing a user’s

request or information need.’’ Based on this definition, a unified pattern-based approach is proposed for nov-elty detection. Queries or topics are automatically classified into specific topics and general topics. Multipleeffective query-related information patterns, namely sentence lengths, combinations of NEs, and opinion pat-terns, are identified and considered at both the retrieval step and the novelty detection step. Three sets ofexperiments were carried out on the data from the TREC novelty tracks 2002–2004, for specific topics, forgeneral topics, and for all the topics. The experimental results show that the proposed approach achieves bet-ter performance at top ranks than the four baseline approaches on topics from the novelty tracks of all thethree years.

The proposed pattern-based approach opens up some further research issues. Even though we have signif-icantly improved the performance of novelty detection for both specific and general topics by using the pro-posed sentence level information patterns, the novelty detection precision for specific topics is much higher.Therefore, our future work will focus on further improving the performance for general topics.

First, the proposed three information patterns only capture part of the characteristics of required answers.Other information patterns, such as the document creation times that we have used in the time-based languagemodel (Li & Croft, 2003), could be helpful in further improving the performance of novelty detection for alltopics. Opinion patterns are only used for opinion topics. Finding similar patterns for event topics could fur-ther improve the performance for event topics. An event pattern such as ‘‘reported by XXX’’ where ‘‘XXX’’could be a person’s name or a news agency, might indicate the description of an event is included in a sentence.Some specific named entity combinations such as a specific time, date and location in a sentence could beanother type of event pattern.

Second, exploring the possibilities of turning more topics into multiple specific questions will be of greatinterest. Currently, only NE-questions that require named entities for answers are considered. However, atopic may be not completely covered by the NE-questions that are automatically transformed at the stepof query analysis. For more topic coverage, other types of question should be considered in addition toNE-questions. Therefore, one direction of our future research is to explore other types of question and dis-cover the related patterns that may indicate the existence of potential answers. One type of question that couldbe considered is the ‘‘Why’’ question that usually asks for the cause of an event or the reason of an opinion.For this type of question, the occurrence of ‘‘because’’, ‘‘the reason is that’’, ‘‘due to’’, or ‘‘caused by’’ in asentence may indicate possible answers. These words or phrases could be used as useful patterns for identify-ing relevant sentences. There are other types of questions that could be considered, such as definition questions(‘‘What’’ questions) and task questions (‘‘How To’’ questions). Many other types of questions have been stud-ied in the question answering research community. A close monitoring of the development of question answer-ing techniques and integrating them into the proposed pattern-based approach will further improve theperformance of novelty detection.

9. Uncited references

Eichmann et al. (2003), Lavrenko and Bruce Croft (2001), Li and Croft (2001), Murdock and Croft (2005),Ponte (1998), Voorhees (2000) and Voorhees (2002).

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 29: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

1086

1087

1088

1089

1090

1091

1092

1093

1094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142

X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx 29

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

UN

CO

RR

EC

TED

PR

OO

F

Acknowledgements

This paper is based on the PhD thesis work of the first author (Li, 2006) at the University of Massachusetts,Amherst. The work was supported in part by the Center for Intelligent Information Retrieval, by SPAW-ARSYSCEN-SD grant numbers N66001-99-1-8912 and N66001-1-8903, and in part by the Defense AdvancedResearch Projects Agency (DARPA) under contract number HR0011-06-C-0023. Any opinions, findings andconclusions or recommendations expressed in this material are the authors’ and do not necessarily reflect thoseof the sponsors.

References

Allan, J., Gupta, R., & Khandelwal, V. (2001). Temporal summaries of news topics. In Proceedings of the 24th annual international ACM

SIGIR conference on research and development in information retrieval (pp. 10–18).Allan, J., Lavrenko, V., & Jin, H. (2000). First story detection in TDT is hard. In Proceedings of the eleventh international conference on

information and knowledge management (pp. 374–381).Allan, J., Paka, R., & Lavrenko, V. (1998). On-line new event detection and tracking. In Proceedings of the 21st annual international ACM

SIGIR conference on research and development in information retrieval (pp. 37–45).Allan, J., Wade, C., & Bolivar, A. (2003). Retrieval and novelty detection at the sentence level. In Proceedings of the 26th annual

international ACM SIGIR conference on research and development in information retrieval (pp. 314–321).Bikel, D. M., Schwartz, R. L., & Weischedel, R. M. (1999). An algorithm that learns what’s in a name. Machine Leaning, 3, 221–

231.Brants, T., Chen, F., & Farahat, A. (2003). A system for new event detection. In Proceedings of the 26th annual international ACM SIGIR

conference on research and development in information retrieval (pp. 330–337).Carbonell, J., & Goldstein, J. (1998). The use of MMR, diversity-based re-ranking for reordering documents and producing summaries. In

Proceedings of the 21st annual international ACM SIGIR conference on research and development in information retrieval (pp. 335–336).Dai, W., & Srihari, R. (2005). Minimal document set retrieval. In Proceedings of the 14th international conference on information and

knowledge management (pp. 752–759).Eichmann, D., & Srinivasan, P. (2002). Novel results and some answers, the University of Iowa TREC-11 results. In NIST Special

Publication: SP 500-251, the eleventh text retrieval conference. http://trec.nist.gov/pubs/trec11/papers/uiowa.eichmann.pdf.Eichmann, D., Srinivasan, P., Light, M., Wang, H., Qiu, X. Y., Arens, R. J. et al. (2003). Experiments in novelty, genes and questions at

the University of Iowa. In NIST Special Publication: SP 500-255, the twelfth text retrieval conference. http://trec.nist.gov/pubs/trec12/papers/uiowa.novelty.genomics.qa.pdf.

Franz, M., Ittycheriah, A., McCarley, J. S., & Ward, T. (2001). First story detection, combining similarity and novelty based approach. In2001 Topic Detection and Tracking (TDT) Workshop Report.

Harman, D. (2002). Overview of the TREC 2002 novelty track. In NIST Special Publication: SP 500-251, the eleventh text retrieval

conference. http://trec.nist.gov/pubs/trec11/papers/NOVELTY.OVER.pdf.Jin, Q., Zhao, J., & Xu, B. (2003). NLPR at TREC 2003: novelty and robust. In NIST Special Publication: SP 500-255, the twelfth text

retrieval conference http://trec.nist.gov/pubs/trec12/papers/chinese-acad-sci.novelty.robust.pdf.Kazawa, H., Hirao, T., Isozaki, H., & Maeda, E. (2003). A machine learning approach for QA and novelty tracks: NTT system

description. In NIST Special Publication: SP 500-255, the twelfth text retrieval conference. (pp. 460–467).Kim, S.-N., Ravichandran, D., & Hovy, E. (2004). ISI novelty track system for TREC 2004. In NIST Special Publication: SP 500-261, the

thirteenth text retrieval conference. http://trec.nist.gov/pubs/trec13/papers/usc-isi.novelty.pdf.Krovetz, R. (1993). Viewing morphology as an inference process, In Proceedings the 16th annual international ACM SIGIR conference on

research and development in information retrieval (pp. 191–202).Kumaran, G., & Allan, J. (2004). Text classification and named entities for new event detection. In Proceedings of the 27th annual

international ACM SIGIR conference on research and development in information retrieval (pp. 297–304).Kumaran, G., & Allan, J. (2006). Simple questions to improve pseudo-relevance feedback results. In The 29th annual international ACM

SIGIR conference on research and development in information retrieval (pp. 661–662).Kwok, K. L., Deng, P., Dinstl, N., Chan, M. (2002). TREC2002, novelty and filtering track experiments using PRICS. In NIST Special

Publication: SP 500-251, the eleventh text retrieval conference. http://trec.nist.gov/pubs/trec11/papers/queens.kwok.pdf.Lavrenko, V., & Bruce Croft, W. (2001). Relevance-based language models. In The 24th annual international ACM SIGIR conference on

research and development in information retrieval (pp. 10–18).Lemur (2006). Lemur toolkit for language modeling and information retrieval. The LEMUR PROJECT by CMU and UMASS. http://

www.lemurproject.org.Li, X. (2003). Syntactic features in question answering. In Proceedings of the 26th annual international ACM SIGIR conference on research

and development in information retrieval (pp. 383–384).Li, X. (2006). Sentence level information patterns for novelty detection, PhD thesis, Department of Computer Science, University of

Massachusetts at Amherst. http://www.mtholyoke.edu/xli/Xiaoyan-PhD-Thesis-2006-Final.pdf.Li, X., & Croft, W. B. (2001). Evaluating question answering techniques in Chinese. In Proceedings of human language technology

conference (pp. 96–101).

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013

Page 30: An information-pattern-based approach to novelty detectionxli/IPM_1157.pdf · In this paper, we investigate if identifying query-related named entity (NEs) 85 patterns in sentences

11431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186

1187

30 X. Li, W.B. Croft / Information Processing and Management xxx (2007) xxx–xxx

IPM 1157 No. of Pages 30, Model 3+

24 October 2007; Disk UsedARTICLE IN PRESS

OR

REC

TED

PR

OO

F

Li, X., & Croft, W. B. (2003). Time-based language models. In Proceedings of the 12th international conference on information and

knowledge management (pp. 469–475).Li, X., & Croft, W. B. (2005). Novelty detection based on sentence level information patterns. In Proceedings of the 14th international

conference on information and knowledge management (pp. 744–751).Li, X., & Croft, W. B. (2006). Improving novelty detection for general topics using sentence level information patterns. In Proceedings of

the 15th international conference on information and knowledge management (pp. 238–247).Litkowski, K. C. (2003). Use of metadata for question answering and novelty tasks. In NIST Special Publication: SP 500-255, the twelfth

text retrieval conference. http://trec.nist.gov/pubs/trec12/papers/clresearch.qa.novelty.pdf.Murdock, V., & Croft, W. B. (2005). A translation model for sentence retrieval. In Proceedings of the conference on human language

technologies and empirical methods in natural language processing (HLT/EMNLP) (pp. 684–691).Ponte, J. (1998). A language modeling approach to information retrieval. PhD thesis, Department of Computer Science, University of

Massachusetts at Amherst.Qi, H., Otterbacher, J., Winkel, A., & Radev, D. T. (2002). The University of Michigan at TREC2002: question answering and novelty

tracks. In NIST Special Publication: SP 500-251, the eleventh text retrieval conference. http://trec.nist.gov/pubs/trec11/papers/umichigan.radev.pdf.

Salton, G., & McGill, M. J. (1983). Introduction to modern information retrieval. McGraw-Hill, ISBN: 007054484.Soboroff, I. (2004). Overview of the TREC 2004 novelty track. In NIST Special Publication: SP 500-261, the thirteenth text retrieval

conference. http://trec.nist.gov/pubs/trec13/papers/NOVELTY.OVERVIEW.pdf.Soboroff, I., & Harman, D. (2003). Overview of the TREC 2003 novelty track. In NIST Special Publication: SP 500-255, the twelfth text

retrieval conference. http://trec.nist.gov/pubs/trec12/papers/NOVELTY.OVERVIEW.pdf.Stokes, N., & Carthy, J. (2001). First story detection using a composite document representation. In Proceedings of human language

technology conference (pp. 134–141).Sun, J., Yang, J., Pan, W., Zhang, H., Wang, B. & Cheng, X. (2003). TREC-2003 novelty and web track at ICT. In NIST Special

Publication: SP 500-255, the twelfth text retrieval conference. http://trec.nist.gov/pubs/trec12/papers/chinese-acad-sci.novelty.web.pdf.Tsai, M., Hsu, M., & Chen, H. (2003). Approach of information retrieval with reference corpus to novelty detection. In NIST Special

Publication: SP 500-255, the twelfth text retrieval conference. http://trec.nist.gov/pubs/trec12/papers/ntu.novelty.pdf.Voorhees, E. M. (2000). Variations in relevance judgments and the measurement of retrieval effectiveness. Information Processing and

Management, 36(5), 697–716.Voorhees, E. M. (2002). Overview of the TREC 2002 question answering track. In NIST Special Publication: SP 500-251, the eleventh text

retrieval conference. http://trec.nist.gov/pubs/trec11/papers/QA11.pdf.Yang, Y., Pierce, T., & Carbonell, J. (1998). A study on retrospective and on-line event detection. In The 21st annual international ACM

SIGIR conference on research and development in information retrieval (pp. 28–36).Yang, Y., Zhang, J. Carbonell, J., & Jin, C. (2002). Topic-conditioned novelty detection. In Proceedings of the eighth ACM SIGKDD

international conference on knowledge discovery and data mining (pp. 688–693).Zhai, C. (2001). Notes on the Lemur TFIDF model, School of Computer Science, CMU www.cs.cmu.edu/lemur/1.9/tfidf.ps.Zhai, C., Cohen, W. W., & Lafferty, J. (2003). Beyond independent relevance: method and evaluation metrics for subtopic retrieval. In

Proceedings of the 26th annual international ACM SIGIR conference on research and development in information retrieval (pp. 10–17).Zhang, Y., Callan, J., & Minka, T. (2002). Novelty and redundancy detection in adaptive filtering. In The 25th annual international ACM

SIGIR conference on research and development in information retrieval (pp.81–88).Zhang, M., Lin, C., Liu, Y., Zhao, L., & Ma, S. (2003). THUIR at REEC 2003: novelty, robust and web. In NIST Special Publication: SP

500-255, the twelfth text retrieval conference. http://trec.nist.gov/pubs/trec12/papers/tsinghuau.hard.pdf.Zhang, M., Song, R., Lin, C., Ma, S., Jiang, Z., Jin, Y. et al. (2002). Expansion-based technologies in finding relevant and new

information: THU TREC2002 novelty track experiments. In NIST Special Publication: SP 500-251, the eleventh text retrieval

conference. http://trec.nist.gov/pubs/trec11/papers/tsinghuau.novelty2.pdf.

UN

C

Please cite this article in press as: Li, X., & Croft, W. B. , An information-pattern-based approach to novelty detection,Information Processing and Management (2007), doi:10.1016/j.ipm.2007.09.013