JOURNAL ARTICLE

Evaluation of Pseudo-Relevance Feedback using Wikipedia

Abstract

Users have specific information needs which are expressed in short queries to information retrieval systems. The queries are unstructured, and they tend to be short and ambiguous in most cases. Using the shallow language statistics including probabilistic or language models such as BM25 or Indri respectively can enhance the retrieval system metrics like Mean Average Precision (MAP). However, such methods depend on query terms and their presence in the retrieved document to define relevance. Query expansion is a technique that can be used to overcome this problem by expanding the query with terms from an initial top few relevant documents. The question that we try to answer is whether the quality of the corpus used for expansion produce a significant improvement MAP and precision at top 30 retrieved documents. We show that the quality and the selection criteria of expansion documents are important factors in query expansion performance.

Keywords:
Computer science Query expansion Relevance (law) Information retrieval Relevance feedback Selection (genetic algorithm) Probabilistic logic Query language Language model Quality (philosophy) Web search query Web query classification Query optimization Sargable Ranking (information retrieval) Search engine Natural language processing Artificial intelligence Image retrieval

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
12
Refs
0.09
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Information Retrieval and Search Behavior
Physical Sciences →  Computer Science →  Information Systems
Web Data Mining and Analysis
Physical Sciences →  Computer Science →  Information Systems
© 2026 ScienceGate Book Chapters — All rights reserved.