Workload sampling for enterprise search evaluation

Tom Rowlands*, David Hawking, Ramesh Sankaranarayana

*Corresponding author for this work

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    3 Citations (Scopus)

    Abstract

    In real world use of test collection methods, it is essential that the query test set be representative of the work load expected in the actual application. Using a random sample of queries from a media company's query log as a 'gold standard' test set we demonstrate that biases in sitemap-derived and top n query sets can lead to significant perturbations in engine rankings and big differences in estimated performance levels.

    Original languageEnglish
    Title of host publicationProceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR'07
    Pages887-888
    Number of pages2
    DOIs
    Publication statusPublished - 2007
    Event30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR'07 - Amsterdam, Netherlands
    Duration: 23 Jul 200727 Jul 2007

    Publication series

    NameProceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR'07

    Conference

    Conference30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR'07
    Country/TerritoryNetherlands
    CityAmsterdam
    Period23/07/0727/07/07

    Fingerprint

    Dive into the research topics of 'Workload sampling for enterprise search evaluation'. Together they form a unique fingerprint.

    Cite this