google example query
play

Google example query: heat in query doesnt match with thermodynamics - PowerPoint PPT Presentation

Overview Introduction to Information Retrieval Recap http://informationretrieval.org 1 IIR 9: Relevance Feedback & Query Expansion Relevance feedback: Basics 2 Hinrich Sch utze Relevance feedback: Details 3 Institute for Natural


  1. Overview Introduction to Information Retrieval Recap http://informationretrieval.org 1 IIR 9: Relevance Feedback & Query Expansion Relevance feedback: Basics 2 Hinrich Sch¨ utze Relevance feedback: Details 3 Institute for Natural Language Processing, Universit¨ at Stuttgart 2008.06.03 4 Global query expansion 1 / 57 2 / 57 Plan for this lecture First: Recap Main topic today: How can we improve recall in search? “aircraft” in query doesn’t match with “plane” in document Google example query: “heat” in query doesn’t match with “thermodynamics” in ˜hospital -hospital -hospitals document Options for improving recall Local methods: Do a “local”, on-demand analysis for a user query Main local method: relevance feedback Global methods: Do a global analysis once (e.g., of collection) to produce thesaurus Use thesaurus for query expansion 3 / 57 4 / 57

  2. Overview Outline Recap Recap 1 1 Relevance feedback: Basics Relevance feedback: Basics 2 2 Relevance feedback: Details Relevance feedback: Details 3 3 4 Global query expansion 4 Global query expansion 5 / 57 6 / 57 Relevance Relevance: query vs. information need The notion of “relevance to the query” is very problematic. Information need i : You are looking for information on We will evaluate the quality of an information retrieval system whether drinking red wine is more effective at reducing your and, in particular, its ranking algorithm with respect to risk of heart attacks than white wine. relevance. Query q : wine and red and white and heart and A document is relevant if it gives the user the information she attack was looking for. Consider document d ′ : He then launched into the heart of his To evaluate relevance, we need an evaluation benchmark with speech and attacked the wine industry lobby for downplaying three elements: the role of red and white wine in drunk driving. A benchmark document collection d ′ is relevant to the query q , but d ′ is not relevant to the A benchmark suite of queries information need i . An assessment of the relevance of each query-document pair User happiness/satisfaction (i.e., how well our ranking algorithm works) can only be measured by relevance to information needs, not by relevance to queries. 7 / 57 8 / 57

  3. Precision and recall A combined measure: F Precision ( P ) is the fraction of retrieved documents that are relevant F allows us to trade off precision against recall. Precision = #(relevant items retrieved) = P (relevant | retrieved) Balanced F : #(retrieved items) F 1 = 2 PR P + R Recall ( R ) is the fraction of relevant documents that are retrieved This is a kind of soft minimum of precision and recall. Recall = #(relevant items retrieved) = P (retrieved | relevant) #(relevant items) 9 / 57 10 / 57 Averaged 11-point precision/recall graph Outline 1 0.8 Precision 0.6 1 Recap 0.4 0.2 Relevance feedback: Basics 2 0 0 0.2 0.4 0.6 0.8 1 Recall Relevance feedback: Details This curve is typical of performance levels for the TREC benchmark. 3 70% chance of getting the first document right (roughly) When we want to look at at least 50% of all relevant documents, then for each relevant document we find, we will have to look at about two nonrelevant 4 Global query expansion documents. That’s not very good. High-recall retrieval is an unsolved problem. 11 / 57 12 / 57

  4. Relevance feedback: Basic idea Relevance Feedback: Example User issues a (short, simple) query Search engine returns set of docs User marks some docs as relevant, some as nonrelevant Search engine computes a new representation of information need – better than initial query Search engine runs new query and returns new results New results have (hopefully) better recall. We can iterate this. We will use the term ad hoc retrieval to refer to regular retrieval without relevance feedback. We will now look at three different examples of relevance feedback that highlight different aspects of the process. 13 / 57 14 / 57 Results for initial query User feedback: Select what is relevant 15 / 57 16 / 57

  5. Results after relevance feedback Ad hoc retrieval for query “canine” (1) source: Fernando D´ ıaz 17 / 57 18 / 57 Ad hoc retrieval for query “canine” (2) User feedback: Select what is relevant source: source: Fernando D´ ıaz Fernando D´ ıaz 19 / 57 20 / 57

  6. Results after relevance feedback Results for initial query Initial query: New space satellite applications Results for initial source: query: 1. 0.539, 08/13/91, NASA Hasn’t Scrapped Imaging Spectrom- Fernando D´ ıaz eter 2. 0.533, 07/09/91, NASA Scratches Environment Gear From Satellite Plan 3. 0.528, 04/04/90, Science Panel Backs NASA Satellite Plan, But Urges Launches of Smaller Probes 4. 0.526, 09/09/91, A NASA Satellite Project Accomplishes In- credible Feat: Staying Within Budget 5. 0.525, 07/24/90, Scientist Who Exposed Global Warming Pro- poses Satellites for Climate Research 6. 0.524, 08/22/90, Report Provides Support for the Critics Of Using Big Satellites to Study Climate 7. 0.516, 04/13/87, Arianespace Receives Satellite Launch Pact From Telesat Canada 8. 0.509, 12/02/87, Telecommunications Tale of Two Companies User then marks relevant documents with “+”. 21 / 57 22 / 57 Expanded query after relevance feedback Results for expanded query * 1. 0.513, 07/09/91, NASA Scratches Environment Gear From Satellite Plan * 2. 0.500, 08/13/91, NASA Hasn’t Scrapped Imaging Spectrom- eter 2.074 new 15.106 space 3. 0.493, 08/07/89, When the Pentagon Launches a Secret Satel- lite, Space Sleuths Do Some Spy Work of Their Own 30.816 satellite 5.660 application 4. 0.493, 07/31/89, NASA Uses ‘Warm’ Superconductors For 5.991 nasa 5.196 eos Fast Circuit 4.196 launch 3.972 aster * 5. 0.492, 12/02/87, Telecommunications Tale of Two Companies 3.516 instrument 3.446 arianespace 6. 0.491, 07/09/91, Soviets May Adapt Parts of SS-20 Missile For Commercial Use 3.004 bundespost 2.806 ss 7. 0.490, 07/12/88, Gaping Gap: Pentagon Lags in Race To 2.790 rocket 2.053 scientist Match the Soviets In Rocket Launchers 2.003 broadcast 1.172 earth 8. 0.490, 06/14/90, Rescue of Satellite By Space Agency To Cost 0.836 oil 0.646 measure $90 Million 23 / 57 24 / 57

  7. Outline Key concept for relevance feedback: Centroid The centroid is the center of mass of a set of points. Recap 1 Recall that we represent documents as points in a high-dimensional space. Thus: we can compute centroids of documents. Relevance feedback: Basics 2 Definition: 1 Relevance feedback: Details � 3 � µ ( D ) = � v ( d ) | D | d ∈ D 4 Global query expansion v ( d ) = � where D is a set of documents and � d is the vector we use to represent the document d . 25 / 57 26 / 57 Centroid: Examples Rocchio algorithm ⋄ The Rocchio algorithm implements relevance feedback in the ⋄ vector space model. ⋄ ⋄ Rocchio chooses the query � q opt that maximizes ⋄ � q opt = arg max [sim( � q , D r ) − sim( � q , D nr )] � q ⋄ Closely related to maximum separation between relevant and nonrelevant docs This optimal query vector is: x 1 d j + [ 1 1 x � � � � � � q opt = � d j − d j ] x | D r | | D r | | D nr | x � � � d j ∈ D r d j ∈ D r d j ∈ D nr D r : set of relevant docs; D nr : set of nonrelevant docs 27 / 57 28 / 57

  8. Rocchio algorithm Rocchio illustrated The optimal query vector is: x x � µ R − � µ NR � q opt 1 d j + [ 1 1 � � � � � � � q opt = d j − d j ] x x | D r | | D r | | D nr | x � � � d j ∈ D r d j ∈ D r d j ∈ D nr � µ R x q-opt = centroid-rel - (centroid-rel - centroid-nonrel) µ NR � We move the centroid of the relevant documents by the difference between the two centroids. circles: relevant documents, Xs: nonrelevant documents � µ R : We had to assume | � µ r | = | � µ nr | = 1 for this derivation. centroid of relevant documents � µ R does not separate rele- vant/nonrelevant. � µ NR : centroid of nonrelevant documents � µ R − µ NR : difference vector Add difference vector to � � µ R . . . . . . to get � q opt � q opt separates relevant/nonrelevant perfectly. 29 / 57 30 / 57 Rocchio 1971 algorithm (SMART) Rocchio relevance feedback illustrated Used in practice: q 0 + β 1 1 � � � � � q m = α� d j − γ d j | D r | | D nr | � � d j ∈ D r d j ∈ D nr q m : modified query vector; q 0 : original query vector; D r and D nr : sets of known relevant and nonrelevant documents respectively; α , β , and γ : weights attached to each term New query moves towards relevant documents and away from nonrelevant documents. Tradeoff α vs. β/γ : If we have a lot of judged documents, we want a higher β/γ . Set negative term weights to 0. Questions? “Negative weight” for a term doesn’t make sense in the vector space model. 31 / 57 32 / 57

Recommend


More recommend