TY - GEN
T1 - Robust relevance-based language models
AU - Li, Xiaoyan
PY - 2006
Y1 - 2006
N2 - We propose a new robust relevance model that can be applied to both pseudo feedback and true relevance feedback in the language-modeling framework for document retrieval. There are three main differences between our new relevance model and the Lavrenko-Croft relevance model. The proposed model brings back the original query into the relevance model by treating it as a short, special document, in addition to a number of top ranked documents returned from the first round retrieval for pseudo feedback, or a number of relevant documents for true relevance feedback. Second, instead of using a uniform prior as in the original relevance model, documents are assigned with different priors according to their lengths (in terms) and ranks in the first round retrieval. Third, the probability of a term in the relevance model is further adjusted by its probability in a background language model. In both cases, we have compared the performance of our model to that of the two baselines: the original relevance model and a linear combination model. Our experimental results show that the proposed new model outperforms both of the two baselines in terms of mean average precision.
AB - We propose a new robust relevance model that can be applied to both pseudo feedback and true relevance feedback in the language-modeling framework for document retrieval. There are three main differences between our new relevance model and the Lavrenko-Croft relevance model. The proposed model brings back the original query into the relevance model by treating it as a short, special document, in addition to a number of top ranked documents returned from the first round retrieval for pseudo feedback, or a number of relevant documents for true relevance feedback. Second, instead of using a uniform prior as in the original relevance model, documents are assigned with different priors according to their lengths (in terms) and ranks in the first round retrieval. Third, the probability of a term in the relevance model is further adjusted by its probability in a background language model. In both cases, we have compared the performance of our model to that of the two baselines: the original relevance model and a linear combination model. Our experimental results show that the proposed new model outperforms both of the two baselines in terms of mean average precision.
KW - Feedback
KW - Language modeling
KW - Query expansion
KW - Relevance models
UR - http://www.scopus.com/inward/record.url?scp=38349173582&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=38349173582&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:38349173582
SN - 9780889866157
T3 - Proceedings of the Fifth IASTED International Conference on Communications, Internet, and Information Technology, CIIT 2006
SP - 341
EP - 348
BT - Proceedings of the Fifth IASTED International Conference on Communications, Internet, and Information Technology, CIIT 2006
T2 - 5th IASTED International Conference on Communications, Internet, and Information Technology, CIIT 2006
Y2 - 29 November 2006 through 1 December 2006
ER -