text retrievalconference (trec)cs626-460-2012/seminar_ppts/... · 2012-04-18 · text...

43
Text REtrieval Conference (TREC) Rekha Ramesh (114388002) Abhishek Mallik (113050019) Aliabbas Petiwala (114380003) 4/13/2012 1 TREC

Upload: others

Post on 28-Jun-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Text REtrieval Conference (TREC)(TREC)

Rekha Ramesh (114388002)

Abhishek Mallik (113050019)

Aliabbas Petiwala (114380003)

4/13/2012 1TREC

Page 2: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Motivation

• You are experimenting in Information Retrieval area

• You have built a retrieval system• You want to test it with realistically sized test data

• You want to test it with realistically sized test data

• You want to compare results across other similar systems

• Where to go?• The answer is TREC ……

4/13/2012 2TREC

Page 3: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

What is TREC ?

• A workshop series that provides the infrastructure for large-scale testing of (text) retrieval technology– Realistic test collections– Realistic test collections

– Uniform, appropriate scoring procedures

– A forum for the exchange of research methodology

4/13/2012 3TREC

Page 4: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

TREC Philosophy

• TREC is a modern example of the Cranfield tradition– system evaluation based on test collection

• Emphasis on advancing the state of the • Emphasis on advancing the state of the art from evaluation results– TREC’s primary purpose is not competitive benchmarking

– Experimental workshop

4/13/2012 4TREC

Page 5: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Yearly Conference Cycle

4/13/2012 5TREC

Page 6: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

A Brief History of TREC

• 1992: first TREC conference– Started by Donna Harman and Charles Wayne as 1 of 3 evaluations in DARPA,s TIPSTER program

– First 3 CDS of documents from this era– First 3 CDS of documents from this era– Open to IR not funded by DARPA

• 25 groups submitted runs

– Two tasks: ad hoc retrieval, routing• 2GB of text, 50 topics• Primarily an exercise in scaling up systems

4/13/2012 6TREC

Page 7: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Adhoc and Routing Task

• Routing task – Same questions are always

being asked, but that new data is being searched. For e.g. News clipping services, library profiling systems.library profiling systems.

• Adhoc task – New questions are being

asked against a static set of data. For e.g. Researcher using a library

4/13/2012 7TREC

Page 8: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Goals of TREC� To encourage research in text retrieval based on large test collections

� To increase communication among industry, academia, and government by creating an open forum for the exchange of research ideasforum for the exchange of research ideas

� To speed the transfer of technology from research labs into commercial products

� To increase the availability of appropriate evaluation techniques for use by industry and academia

4/13/2012 8TREC

Page 9: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

TREC Tracks� TREC-1(1992) and TREC-2(1993) concentrated on ad hoc task and the routing task

�Starting in TREC-3, a variety of other ‘‘tracks,’’ were introduced (shown in table)

� The tracks invigorate TREC by focusing research on new areas or particular aspects of text retrieval

� The tracks represent the majority of the work done in TREC

4/13/2012 9TREC

Page 10: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

TREC Tracks

• Set of tracks in a particular TREC depends on:– Interest of participants

– Appropriateness of task to TREC– Appropriateness of task to TREC

– Needs of sponsors

– resource constraints

• Need to submit proposal for new track in writing to NIST

4/13/2012 10TREC

Page 11: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Tasks Performed in TREC

4/13/2012 11TREC

Page 12: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Use of only static text

• Million Query Track: – to explore ad hoc retrieval on a large document set.

• Robust Track: – to focus research on improving the consistency of retrieval technology

by concentrating on poorly performing topics. by concentrating on poorly performing topics.

Use of streamed text• Filtering:

– for each document in a document stream, decide whether to retrieve it in response to standing query

– Routing, batch filtering, adaptive filtering

4/13/2012 12TREC

Page 13: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Human in the LOOP

• Interactive: Investigate searching as an interactive task by examining the process as well as the outcome– Observational study of subjects using live web to perform

search task– Observational study of subjects using live web to perform

search task

– Controlled laboratory experiment of hypothesis suggested by observations

• HARD: High Accuracy Retrieval from Documents– to improve retrieval performance by targeting retrieval

results to the specific user

4/13/2012 13TREC

Page 14: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Beyond just English

• CLIR: Cross Lingual Information Retrieval– documents are in one language and topics are in another language

Multilingual (Spanish & Chinese):• Multilingual (Spanish & Chinese):– Investigates the retrieval performance when both documents and topics is in a language other than English

4/13/2012 14TREC

Page 15: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Beyond text

• Confusion Track: investigates how retrieval performance is affected through noisy and confused data– Data processed through OCR

• SDR Track: retrieval methodology for spoken documents

• The Video Track: to promote progress in content based retrieval task

4/13/2012 15TREC

Page 16: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Web Searching

• Web Track: build a test system that more closely mimics the retrieval environment of WWW– Documents were a collection of web pages

• Terabyte Track: develop an evaluation methodology for terabyte-scale document collections.methodology for terabyte-scale document collections.– This collection contains a large proportion of the crawlable

pages in .gov, including html and text, plus extracted text of pdf, word and postscript files

• Enterprise Track: satisfying a user who is searching the data of an organization to complete some task.– Enterprise data generally consists of diverse types such as

published reports, intranet web sites, and email, 4/13/2012 16TREC

Page 17: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Answers, not Docs

• Novelty Track: investigate systems’ abilities to locate relevant and new (non redundant) information within an ordered set of documents.

• Q& A Track: systems that return actual answers, as opposed to ranked lists of documents, in response to a question.as opposed to ranked lists of documents, in response to a question.– the test set of questions consisted of factoid, list, and

definition questions.

DEFINITION What is a comet?FACTOID How often does it approach the earth?LIST In what countries was the comet visible on its

last return?OTHER

4/13/2012 17TREC

Page 18: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Retrieval in Domains

• Legal Track : to focus specifically on the problem of e-discovery, the effective production of digital or digitized documents as evidence in litigation

• Genome Track: a forum for evaluation of information access systems in the genomics domain – document collections were a set of full-text articles from

several biomedical journals

4/13/2012 18TREC

Page 19: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Personal Documents

• Blogs Track: explore information seeking behavior in the blogosphere– to discover the similarities and differences between blog

search and other types of search

• Spams Track: to evaluate how well systems are able to separate spam and ham (non-spam) when given an email sequence– The task involved classifying email messages as ham or spam,

differing in the amount and frequency of the feedback the system received.

4/13/2012 19TREC

Page 20: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Participation in TREC

4/13/2012 20TREC

Page 21: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

TREC Impacts

• Test collections– Papers in general literature use TREC collections

• Incubator for new research areas• Incubator for new research areas– PhD theses resulting from CLIR, SDR, QA

• Common evaluation methodology and improved measures for text retrieval– Document best practices in IR research methodology for new researchers

4/13/2012 21TREC

Page 22: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

TREC Impacts

• Open forum for exchange of research– TREC papers figure prominently in IR syllabi on the web

– Publication of all results prevents unsuccessful – Publication of all results prevents unsuccessful research from being duplicated

• Technology transfer– Impact is far greater than just those who actually participate

4/13/2012 22TREC

Page 23: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

TREC Approach

4/13/2012 23TREC

Page 24: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Creating a test collection for Adhoc task

4/13/2012 24TREC

Page 25: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Documents

• Must be representative of real task of interest– Genre– Diversity (subject, style and vocabulary)– Amount– Amount– Full text vs. abstract

• TREC– Generally newswire/news paper– General interest topics– Full text

4/13/2012 25TREC

Page 26: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Topics

• Distinguish between statement of user need (topic) & system data structure (query)– Topics gives criteria for relevance– Allows for different query construction techniques

• TREC topics are NOT all created equal– 1-150: very detailed, rich content– 1-150: very detailed, rich content– 151-250: method of topic creation resulted in focused and easy topics– 201-250: single sentence only– 301-450: title is set of hand-picked keywords

• A topic generally consists of four sections:– an identifier, – a title (consist of up to three words that best describe the topic), – a description (one sentence description of the topic area), and – a narrative (a concise description of what makes a document relevant)

4/13/2012 26TREC

Page 27: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

An Example Topic

<num> Number: 951

<title> Mutual Funds

<desc> Description: Blogs about mutual funds performance and trends.

<narr> Narrative: Ratings from other known sources <narr> Narrative: Ratings from other known sources (Morningstar) correlative to key performance indicators (KPI) such as inflation, currency markets and domestic and international vertical market outlooks. News about mutual funds, mutual fund managers and investment companies. Specific recommendations should have supporting evidence or facts linked from known news or corporate sources. (Not investment spam or pure, uninformed conjecture.)4/13/2012 27TREC

Page 28: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Relevance Judgments (1)

• The relevance judgments turns a set of documents and topics into a test collection

• Retrieve all of the relevant documents and • Retrieve all of the relevant documents and none of the irrelevant documents

• TREC usually uses binary relevance judgments

4/13/2012 28TREC

Page 29: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Relevance Judgments (2)

• In test collections, judgments are usually binary, static and assumed to be complete

• But…• But…– Relevance is idiosyncratic

– Relevance does not entail utility

– Documents have different degrees of relevance

– Relevance can change over time with the same user

– For realistic collections judgments cannot be complete

4/13/2012 29TREC

Page 30: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Relevance Judgments (3)

• Consistency• Idiosyncratic nature of relevance judgments does not effect comparative results

• Incompleteness• Complete judgments should be unbiased

• TREC pooling has been adequate to produce unbiased judgments

4/13/2012 30TREC

Page 31: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Creating relevance judgments

4/13/2012 31TREC

Page 32: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Evaluation at TREC

• In TREC, ad hoc tasks are evaluated using the trec-evalpackage

• This package reports about 85 different numbers for a runrun

• Including recall and precision at various cut-off levels

• And single valued summary measures that are derived from recall and precision.

4/13/2012 32TREC

Page 33: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Evaluation Measure Criteria

• Related to user satisfaction

• Interpretable

• Able to average or collect• Able to average or collect

• Have high discrimination power

• Able to be analyzed

4/13/2012 33TREC

Page 34: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Ranked Retrieval Chart

4/13/2012 34TREC

Page 35: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Evaluation Contingency Table

Relevant Non Relevant

Retrieved r n-r

Non Retrieved R-r N-n-R+r

N - Number of docs in collectionn - Number of docs retrievedR - Number of docs relevantr - Number of relevant docs retrieved

4/13/2012 35TREC

Page 36: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Recall Precision Graph

4/13/2012 36TREC

Page 37: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Uninterpolated R-P curve for Single Topic

4/13/2012 37TREC

Page 38: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Interpolated R-P curve for Single Topic

4/13/2012 38TREC

Page 39: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Single Number Summary Score• Precision (n) : r / n

• Recall (n) : r / R

• F-measure: 2PR/(P+R)

Average: Avg ( Prec (rank of rd))• Average: Avg rd( Prec (rank of rd))

• Precision at K: Precision at fixed low levels of retrieved result

• R-Precision: r/|R| where r is no. of relevant documents from the top |R| documents of a system.

• Rank of first relevant (expected search length)

4/13/2012 39TREC

Page 40: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Average Precision

• Advantage– Sensitive to entire ranking: changing a rank will change final score

– Stable: a small change in ranking makes relatively small change in scoresrelatively small change in scores

– Has both precision and recall oriented factors• Ranks closest to 1 receives largest weight• computed over all relevant documents

• Disadvantage– Less easily interpretable

4/13/2012 40TREC

Page 41: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

Summary

• TREC emphasizes individual experiments evaluated against benchmark task– Leverages modest government investments into substantially more R&D than could Leverages modest government investments into substantially more R&D than could funded directly

– Improves state-of-the-art

– Accelerates technology transfer

4/13/2012 41TREC

Page 42: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

References (1)• Harman D. (Ed.). (1993). The First Text REtrieval Conference (TREC-1). National

Institute of Standards and Technology Special Publication 500-207, Gaithersburg, Md. 20899

• Donna Harman. Overview of the fourth Text REtrieval Conference (TREC-4). In D. K. Harman, editor, Proceedings of the Fourth Text REtrieval Conference (TREC-4), pages 1–23, October 1996. NIST Special Publication 500-236.

• Ellen M. Voorhees. Overview of the sixteenth Text REtrieval Conference (TREC-16). In Ellen M. Voorhees, editor, Proceedings of the Sixteenth Text REtrievalConference (TREC-4), pages 1–16, November 6–9, 2007. NIST Special Publication

• All subsequent 19 Text REtrieval Conference (TREC) publications

• Ellen M. Voorhees. Variations in relevance judgments and the measurement of retrieval effectiveness. Information Processing and Management, 36:697–716, 2000

4/13/2012 42TREC

Page 43: Text REtrievalConference (TREC)cs626-460-2012/seminar_ppts/... · 2012-04-18 · Text REtrievalConference (TREC) RekhaRamesh(114388002) AbhishekMallik(113050019) AliabbasPetiwala(114380003)

References (2)

• Chris Buckley, Darrin Dimmick, Ian Soboroff, and Ellen Voorhees. Bias and the limits of pooling for large collections. Information Retrieval, 10:491–508, 2007.

• Chris Buckley. Trec-eval IR evaluation package. Available from http:/trec.nist.gov/trec-eval/

• Chris Buckley, Ellen M. Voorhees. Retrieval Evaluation with Incomplete • Chris Buckley, Ellen M. Voorhees. Retrieval Evaluation with Incomplete Information.SIGIR’04, July 25–29, 2004, Sheffield, South Yorkshire, UK.

4/13/2012 43TREC