• No results found

Information Retrieval. Lecture 8 - Relevance feedback and query expansion. Introduction. Overview. About Relevance Feedback. Wintersemester 2007

N/A
N/A
Protected

Academic year: 2021

Share "Information Retrieval. Lecture 8 - Relevance feedback and query expansion. Introduction. Overview. About Relevance Feedback. Wintersemester 2007"

Copied!
8
0
0

Loading.... (view fulltext now)

Full text

(1)

Information Retrieval

Lecture 8 - Relevance feedback and query

expansion

Seminar f¨ur Sprachwissenschaft International Studies in Computational Linguistics

Wintersemester 2007

1 / 32

Introduction

◮An information need may be expressed using different

keywords (synonymy)

→impact on recall

→examples: ship vs boat, aircraft vs airplane

◮Solutions: refining queries manuallyorexpanding queries

(semi) automatically

◮Semi-automatic query expansion:

•local methods based on the retrieved documents and the query (ex: Relevance Feedback)

•global methods independent of the query and results (ex: thesaurus, spelling

corrections)

2 / 32

Overview

About Relevance Feedback The Rocchio algorithm

About probabilistic relevance feedback When to use Relevance Feedback Relevance Feedback and the web

Evaluation of Relevance Feedback strategies Other local methods for query expansion Global methods for query expansion

3 / 32

About Relevance Feedback

About Relevance Feedback

Feedback given by the user about the relevance of the documents in the initial set of results

(2)

About Relevance Feedback

About Relevance Feedback (continued)

◮Based on the idea that:

(i) defining good queries is difficult when the collection is (partly) unknown

(ii) judging particular documents is easy

◮Allows to deal with situations where the user’s information

needs evolve with the checking of the retrieved documents

◮Example: image search engine

http://nayana.ece.ucsb.edu/imsearch/imsearch.html

5 / 32

About Relevance Feedback

Relevance feedback example 1

6 / 32

About Relevance Feedback

Relevance feedback example 1

7 / 32

About Relevance Feedback

Relevance feedback example 1

(3)

About Relevance Feedback

Relevance feedback example 2

Query: New space satellite applications

+ 1. 0.539, 08/13/91, NASA Hasn’t Scrapped Imaging Spectrometer

+ 2. 0.533, 07/09/91, NASA Scratches Environment Gear From Satellite Plan

3. 0.528, 04/04/90, Science Panel Backs NASA Satellite Plan, But Urges Launches of Smaller Probes

4. 0.526, 09/09/91, A NASA Satellite Project Accomplishes Incredible Feat: Staying Within Budget

5. 0.525, 07/24/90, Scientist Who Exposed Global Warming Proposes Satellites for Climate Research

6. 0.524, 08/22/90, Report Provides Support for the Critics Of Using Big Satellites to Study Climate

7. 0.516, 04/13/87, Arianespace Receives Satellite Launch Pact From Telesat Canada

+ 8. 0.509, 12/02/87, Telecommunications Tale of Two Companies

9 / 32

About Relevance Feedback

Relevance feedback example 2

2.074 new 15.106 space 30.816 satellite 5.660 application 5.991 nasa 5.196 eos 4.196 launch 3.972 aster 3.516 instrument 3.446 arianespace 3.004 bundespost 2.806 ss 2.790 rocket 2.053 scientist 2.003 broadcast 1.172 earth 0.836 oil 0.646 measure 10 / 32

About Relevance Feedback

Relevance feedback example 2

+ 1. 0.513, 07/09/91, NASA Scratches Environment Gear From Satellite Plan

+ 2. 0.500, 08/13/91, NASA Hasn’t Scrapped Imaging Spectrometer

3. 0.493, 08/07/89, When the Pentagon Launches a Secret Satellite, Space Sleuths Do Some Spy Work of Their Own 4. 0.493, 07/31/89, NASA Uses ’Warm‘ Superconductors For Fast Circuit

+ 5. 0.492, 12/02/87, Telecommunications Tale of Two Companies

6. 0.491, 07/09/91, Soviets May Adapt Parts of SS-20 Missile For Commercial Use

7. 0.490, 07/12/88, Gaping Gap: Pentagon Lags in Race To Match the Soviets In Rocket Launchers

8. 0.490, 06/14/90, Rescue of Satellite By Space Agency To Cost \$90 Million

11 / 32

The Rocchio algorithm

The Rocchio algorithm

◮Standard algorithm for relevance feedback (SMART, 70s) ◮Integrates a measure of relevance feedback into the Vector

Space Model

◮Idea: we want to find a query vectorq~opt

• maximizing the similarity with relevant documents while

• minimizing the similarity with non-relevant docu-ments ~ qopt=argmax ~ q [sim(~q,Cr)−sim(~q,Cnr)]

With the cosine similarity, this gives:

~ qopt= 1 |Cr| X ~ dj∈Cr ~ dj− 1 |Cnr| X ~ dj∈Cnr ~ dj 12 / 32

(4)

The Rocchio algorithm

The Rocchio algorithm (continued)

◮Problem with the above metrics: the set of relevant

documents is unknown

◮Instead, we produce the modified querym:

~ qm=α ~q0+β 1 |Dr| X ~ dj∈Dr ~ dj−γ 1 |Dnr| X ~ dj∈Dnr ~ dj where: ~

q0 is the original query vector

Dr is the set of known relevant documents

Dnr is the set of known non-relevant documents α, β, γ are balancing weights (judge vs system)

13 / 32

The Rocchio algorithm

The Rocchio algorithm (continued)

Remarks:

◮Negative weights are usually ignored

◮Rocchio-based relevance feedback improves both recall and

precision

◮For reaching high recall, many iterations are needed ◮Empirically determined values for the balancing weights:

α= 1 β= 0.75 γ= 0.15

◮Positive feedback is usually more valuable than negative

feedback:β > γ

14 / 32

The Rocchio algorithm

Rocchio algorithm: exercise

Consider the following collection (one doc per line):

good movie trailer shown trailer with good actor unseen movie

a dictionary made of the wordsmovie,trailerandgood, and an IR system using the standardtf−idf weighting (without normalisation).

Assuming a user judges the first 2 documents relevant for the querymovie trailer. What would be the Rocchio-revised query ?

15 / 32

About probabilistic relevance feedback

About probabilistic relevance feedback

◮Alternative to the Rocchio algorithm, use a document

classification instead of a Vector Space Model-based retrieval P(xt= 1|R= 1) = |VRt| |VR| P(xt= 0|R= 0) = nt− |VRt| N− |VR| where:

N is the total number of documents

nt is the number of documents containing t

VR is the set of known relevant documents

VRt is the set of known relevant documents

con-taining t

◮Problem: no memory of the original query

(5)

When to use Relevance Feedback

When to use Relevance Feedback

◮Relevance Feedback does not work when:

• the query is misspelled

• we want cross-language retrieval

• the vocabulary is ambiguous

→the users do not have sufficient initial knowledge

• the query concerns an instance of a general concept (e.g.felines)

• the documents are gathered into subsets each using a different vocabulary

• the query has disjunctive answer sets (e.g.“the pop star that worked at KFC”)

→there exist several prototypes of relevant documents

◮Practical problem: refining leads to longer queries

that need more time to process

17 / 32

Relevance Feedback and the web

Relevance Feedback and the web

◮Few web IR systems use relevance feedback

• hard to explain to users

• users are mainly interested in fast retrieval (i.e. no iterations)

• users usually are not interested in high recall

◮Nowadays: clickstream-based feedback (which links are

clicked on by users)

→implicit feedback from the writer rather than feedback from the reader

18 / 32

Evaluation of Relevance Feedback strategies

Evaluation of Relevance Feedback strategies

◮Note that improvements brought by the relevance feedback

decrease with the number of iterations, usually one round gives good results

◮Several evaluation strategies: (a) comparative evaluation

queryq0→prec/recall graph

queryqm→prec/recall graph

→usually +50% of mean average precision

(partly comes from the fact that known relevant documents are higher ranked)

19 / 32

Evaluation of Relevance Feedback strategies

Evaluation of Relevance Feedback strategies

(continued)

◮Evaluation strategies: (b) residual collection

→same technique as above but by looking at the set of retrieved documents - the set of assessed relevant documents

→the performance measure drops

(c) using two similar collections

collection #1 is used for querying and giving rele-vance feedback

collection #2 is used for comparative evaluation

→q0andqmare compared on collection #2

(6)

Evaluation of Relevance Feedback strategies

Evaluation of Relevance Feedback strategies

(continued)

◮Evaluation strategies: (d) user studies

→e.g.time-based comparison of retrieval, user satisfaction, etc.

→user utility is a fair evaluation as it corresponds to real system usage

21 / 32

Other local methods for query expansion

Overview

About Relevance Feedback The Rocchio algorithm

About probabilistic relevance feedback When to use Relevance Feedback Relevance Feedback and the web

Evaluation of Relevance Feedback strategies Other local methods for query expansion Global methods for query expansion

22 / 32

Other local methods for query expansion

Pseudo Relevance Feedback

◮Akablind relevance feedback

◮No need of an extended interaction between the user and

the system

◮Method:

• normal retrieval to find an initial set of most relevant documents

• assumptionthat the topkdocuments are relevant

• relevance feedback defined accordingly

◮Works with the TREC Ad Hoc task

→lnc.ltc (precision atk= 50): no-RF 62.5 %, RF 72.7 %

◮Problem: distribution of the documents may influence

the results

23 / 32

Other local methods for query expansion

Indirect Relevance Feedback

◮Uses evidences rather than explicit feedback

◮Example: number of clicks on a given retrieved document ◮Not user-specific

◮More suitable for web IR, since it does not need an extra

action from the user

(7)

Global methods for query expansion

Overview

About Relevance Feedback The Rocchio algorithm

About probabilistic relevance feedback When to use Relevance Feedback Relevance Feedback and the web

Evaluation of Relevance Feedback strategies Other local methods for query expansion Global methods for query expansion

25 / 32

Global methods for query expansion

Vocabulary tools for query reformulation

Tools displaying:

◮a list of close terms belonging to the dictionary ◮information about the query words that were omitted (cf

stop-list)

◮the results of stemming → ≈debugging environnement

26 / 32

Global methods for query expansion

Query logs and thesaurus

◮Users select among query suggestions that are built either

fromquery logsorthesaurus

◮Replacement words are extracted from thesaurus according

to their proximity to the initial query word

◮Thesaurus can be developed: • manually (e.g.biomedicine)

• automatically (cf below)

◮NB: query expansion

(i) increases recall

(ii) may need users’ relevance on query terms (6= documents)

27 / 32

Global methods for query expansion

Automatic thesaurus generation

Analyze of the collection for building the thesaurus automatically:

1. Using word co-occurrences (co-occurring words are more likely to belong to the same query field)

→may contain false positives (example: apple)

2. Using a shallow grammatical analyzes to find out relations between words

example:cooked, eaten, digested→food

Note that co-occurrence-based thesaurus are more robust, but grammatical-analyzes thesaurus are more accurate

(8)

Global methods for query expansion

Building a co-occurrence-based thesaurus

◮We build a term-document matrixAwhereA[t,d] =wt,d

(e.g.normalizedtf−idf)

◮We then calculateC=A.AT C=    c11 · · · c1n ... ... ... cm1 · · · cmn   

cijis the similarity score between termsiandj

29 / 32

Global methods for query expansion

Automatically built thesaurus

30 / 32

Global methods for query expansion

Conclusion

◮Query expansion using either local methods: • Rocchio algorithm for Relevance Feedback

• Pseudo Relevance Feedback

• Indirect Relevance Feedback

◮or global ones: • Query logs

• Thesaurus

◮Thesaurus-based query expansion increases recall but may

decrease precision (cfambiguous terms)

◮High cost of thesaurus development and maintenance ◮Thesaurus-based query expansion is less efficient than

Rocchio Relevance Feedback but may be as good as Pseudo Relevance Feedback

31 / 32

Global methods for query expansion

References

• C. Manning, P. Raghavan and H. Sch¨utze, Introduc-tion to InformaIntroduc-tion Retrieval

http://nlp.stanford.edu/IR-book/pdf/ chapter09-queryexpansion.pdf

• Chris Buckley and Gerard Salton and James Allan

The effect of adding relevance information in a rele-vance feedback environment(1994)

http://citeseer.ist.psu.edu/buckley94effect.html • Ian Ruthven and Mounia LalmasA survey on the use

of relevance feedback for information access systems

(2003)

http://personal.cis.strath.ac.uk/

ir/papers/ker.pdf

References

Related documents

However, we still expect positive albeit smaller abnormal returns for our managerial owner- ship portfolios even after industry adjustment, because not all firms in the

Die RolIe der Meiofauna als Nahrung ffir Sipunculiden wurde an Phascolosoma vulgare (DE BLAmVlLLE), Phascolosoma elongatum (K~F~RST~IN) und Physcosoma granulatum

The water quality studies of the San River conducted in 2010–2011 in the area of the San Valley Landscape Park using benthic diatoms as bio-indicators indicated

Pre- RTW Visit RTW med exam Support and follow-up Subacute phase Acute event Sick note.. Entreprise Supervisor, managers Co-workers Worker

LPR treatment requires acid suppression to be aggressive and for a prolonged period of time than GERD in view of the fact that only a small amount of acid reflux into

Work stream 2.2 Education Preventing alcohol and drug use in young people 2.3 Prevention Community focused: positive attitudes and positive choices 2.4 Prevention Early

(1) Whenever the seizure of property by any police officer is reported to a Magistrate under the provisions of this Code, and such property is not produced before a Criminal