class09-im - Three classic approaches to IR 1 Recall:...

Info iconThis preview shows pages 1–8. Sign up to view the full content.

View Full Document Right Arrow Icon
Three “classic” approaches to IR 1 Recall: Boolean Retrieval 1 if play contains word, 0 otherwise Brutus AND Caesar but NOT Calpurnia 2
Background image of page 1

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon
Recall: Vector Space Retrieval 3 Probabilistic IR Chapter 11 Traditional Probabilistic IR model Traditionally: neat ideas, but they’ve never won on performance. Chapter 12 Statistical Language Models Very hot right now 4
Background image of page 2
Why probabilities in IR? User Information Need Documents Document Representation Query Representation How to match? In traditional IR systems, matching between each document and query is attempted in a semantically imprecise space of index terms. Probabilities provide a principled foundation for uncertain reasoning. Can we use probabilities to quantify our uncertainties? Uncertain guess of whether document has relevant content Understanding of user need is uncertain 5 But frst . .. Probability review Independent events Let a, b be two events, with probability P ( a ) and P ( b ). The events a and b are independent if and only if: P ( a ! b ) = P ( a ) P ( b ) In general, a 1 , a 2 , . .. , a n are independent if and only if: P ( a 1 ! a 2 ! ... ! a n ) = P ( a 1 ) P ( a 2 )... P ( a n ) 6
Background image of page 3

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon
Probability review Let a, b be two events, with probability P ( a ) and P ( b ). Conditional probability P ( a | b ) is the probability of a given b, also called the conditional probability of a given b . Conditional independence The events a 1 , . .., a n are conditionally independent if and only if: P ( a i | a j ) = P(a i ) for all i and j. 7 Example Independent a and b are the results of throwing two dice P ( a =5 | b =3) = P ( a =5) = 1 /6 Not independent a and b are the results of throwing two dice t is the sum of the two dice t = a + b P ( t =8 | a =2) = 1 /6 P ( t =8 | a =1) = 0 8
Background image of page 4
Example P ( a ) = x + y P ( b ) = w + x P(a | b) = x / ( w + x ) P ( a | b) P ( b ) = P ( a ! b ) = P ( b | a ) P ( a ) a b w z y x where a is the event not a 9 Bayes theorem Notation Let a, b be two events. P ( a | b ) is the probability of a given b Bayes Theorem P ( a | b ) = Derivation P ( a | b) P ( b ) = P ( a ! b ) = P ( b | a ) P ( a ) P ( b | a ) P ( a ) P ( b ) 10
Background image of page 5

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon
Bayes theorem Terminology used with Bayes Theorem P ( a | b ) = P ( a ) is called the prior probability of a P ( a | b ) is called the posterior probability of a given b P ( b | a ) P ( a ) P ( b ) 11 Example of Bayes theorem Example a Weight over 200 lb. b Height over 6 ft. Over 200 lb Over 6 ft w z y x P ( a | b ) = x / ( w + x ) = x / P ( b ) P ( b | a ) = x / ( x + y ) = x / P ( a ) x is P ( a ! b ) 12
Background image of page 6
IR based on Language Model (LM) query d1 d2 dn Information need document collection generation A common search heuristic is to use words that you expect to fnd in
Background image of page 7

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon
Image of page 8
This is the end of the preview. Sign up to access the rest of the document.

This note was uploaded on 01/21/2011 for the course CSCP 689 taught by Professor James during the Spring '10 term at Texas A&M.

Page1 / 22

class09-im - Three classic approaches to IR 1 Recall:...

This preview shows document pages 1 - 8. Sign up to view the full document.

View Full Document Right Arrow Icon
Ask a homework question - tutors are online