wired week 3 syllabus update (next week) readings overview - quick review of last week’s ir models...

23
WIRED Week 3 WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries Assignment Overview & Scheduling - Leading WIRED Topic Discussions - Web Information Retrieval System Evaluation & Presentation Projects and/or Papers Discussion - Initial Ideas - Evaluation - Revise & Present

Upload: cuthbert-carroll

Post on 04-Jan-2016

214 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

WIRED Week 3WIRED Week 3• Syllabus Update (next week)

• Readings Overview- Quick Review of Last Week’s IR Models (if time)- Evaluating IR Systems- Understanding Queries

• Assignment Overview & Scheduling- Leading WIRED Topic Discussions- Web Information Retrieval System Evaluation &

Presentation

• Projects and/or Papers Discussion- Initial Ideas- Evaluation- Revise & Present

Page 2: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Evaluating IR SystemsEvaluating IR Systems

• Recall and Precision• Alternative Measures• Reference Collections- What- Why

• Trends

Page 3: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Why Evaluate IR Systems?Why Evaluate IR Systems?

• Leave it to the developers?- No bugs- Fully functional

• Let the market (users) decide?- Speed- (Perceived) accuracy

• Relevance is relevant- Different types of searches, data and users

• “How precise is the answer set?” p 73

Page 4: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Retrieval Performance EvaluationRetrieval Performance Evaluation

• Task- Batch or Interactive- Each needs a specific interface

• Setting• Context- New search- Monitoring

• Usability- Lab tests- Real world (search log) analysis

Page 5: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Recall and PrecisionRecall and Precision

• Basic evaluation measurement for IR system performance

• Recall: the fraction of relevant documents retrieved- 100% is perfect recall- Every document that is relevant is found

• Precision: the fraction of retrieved documents which are relevant- 100% relevancy is perfect precision- How good the recall is

Page 6: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Recall and Precisions goalsRecall and Precisions goals

• Everything is found (recall)• The right set of documents is pulled from the

found set (precision)

• What about ranking?- Ranking is an absolute measure of relevance for

the query.- Ranking is Ordinal in almost all cases

Page 7: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Recall and Precision ConsideredRecall and Precision Considered

• 100 documents have been analyzed• 10 documents relevant to the query in the set- 4 documents are found and all are relevant• ??% recall, ??% precision

- 8 documents are found, but 4 are relevant• ??% recall, ??% precision

• Which is more important?

Page 8: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Recall and Precision Appropriate?Recall and Precision Appropriate?

• Disagreements over perfect sets• User errors in using results• Redundancy of results- Result diversity- Metadata

• Dynamic data- Indexable- Recency of information may be key

• A single measure is better- Combinatory- User evaluation

Page 9: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Back to the UserBack to the User

• User evaluation• Is one answer good enough? Rankings• Satisficing

• Studies of Relevance are key

Page 10: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Other Evaluation MeasuresOther Evaluation Measures

• Harmonic Mean- Single, combined measure- Between 0 (none) & 1 (all)- Only high when both P & R are high- Still a percentage

• E measure- User determines (parameter) value of R & P- Different tasks (legal, academic)- An interactive search?

Page 11: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Coverage and NoveltyCoverage and Novelty

• System effects- Relative recall- Relative effort

• sMore natural, user understandable measure• User knows some % documents are relevant• Coverage = % documents user expects• Novelty = % of documents user didn’t know of- Content of document- Document itself- Author of document- Purpose of document

Page 12: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Reference CollectionsReference Collections

• Testbeds for IR evaluation• TREC (Text Retrieval Conference) set- Industry focus- Topic-based or General- Summary tables for tasks (queries)- R & P averages- Document analysis- Measures for each topic

• CACM (general CS)• ISI (academic, indexed, industrial)

Page 13: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Trends in IR EvaluationTrends in IR Evaluation

• Personalization• Dynamic Data• Multimedia• User Modeling• Machine Learning (CPU/$)

Page 14: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Understanding QueriesUnderstanding Queries

• Types of Queries:- Keyword- Context- Boolean- Natural Language

• Pattern Matching- More like this…- Metadata

• Structural Environments

Page 15: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

BooleanBoolean

• AND, OR, NOT• Combination or individually• Decision tree parsing for the system• Not so easy for the user when advanced

queries• Hard to backtrack and see differences in

results

Page 16: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

KeywordKeyword

• Single word (most common)- Sets- “Phrases”

• Context- “Phrases”- Near (# value in characters, words, documents

links)

Page 17: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Natural LanguageNatural Language

• Asking• Quoting• Fuzzy matches• Different evaluation methods might be needed• Dynamic data “indexing” problematic• Multimedia challenges

Page 18: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Pattern MatchingPattern Matching

• Words• Prefixes “comput*”• Suffixes “*ology”• Substrings “*exas*”• Ranges “four ?? years ago”• Regular Expressions (GREP)• Error threshold• User errors

Page 19: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Query ProtocolsQuery Protocols

• HTTP• Z39.50- Client – Server API

• WAIS- Information/ database connection

• ODBC• JDBC• P2P

Page 20: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Assignment Overview & SchedulingAssignment Overview & Scheduling

• Leading WIRED Topic Discussions- # in class = # of weeks left?

• Web Information Retrieval System Evaluation & Presentation- 5 page written evaluation of a Web IR System- technology overview (how it works)- a brief history of the development of this type of

system (why it works better)- intended uses for the system (who, when, why)- (your) examples or case studies of the system in

use and its overall effectiveness

Page 21: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

• How can (Web) IR be better?- Better IR models- Better User Interfaces

• More to find vs. easier to find

• Scriptable applications• New interfaces for applications• New datasets for applications

Projects and/or Papers OverviewProjects and/or Papers Overview

Page 22: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Project Idea #1 – simple HTMLProject Idea #1 – simple HTML

• Graphical Google

• What kind of document?

• When was the document created?

Page 23: WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries

Project IdeasProject Ideas

• Google History: keeps track of what I’ve seen and not seen

• Searching when it counts: Financial and Health information requires guided, quality search