Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/345508.345545acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
Article
Free access

IR evaluation methods for retrieving highly relevant documents

Published: 01 July 2000 Publication History

Abstract

This paper proposes evaluation methods based on the use of non-dichotomous relevance judgements in IR experiments. It is argued that evaluation methods should credit IR methods for their ability to retrieve highly relevant documents. This is desirable from the user point of view in modern large IR environments. The proposed methods are (1) a novel application of P-R curves and average precision computations based on separate recall bases for documents of different degrees of relevance, and (2) two novel measures computing the cumulative gain the user obtains by examining the retrieval result up to a given ranked position. We then demonstrate the use of these evaluation methods in a case study on the effectiveness of query types, based on combinations of query structures and expansion, in retrieving documents of various degrees of relevance. The test was run with a best match retrieval system (In-Query1) in a text database consisting of newspaper articles. The results indicate that the tested strong query structures are most effective in retrieving highly relevant documents. The differences between the query types are practically essential and statistically significant. More generally, the novel evaluation methods and the case demonstrate that non-dichotomous relevance assessments are applicable in IR experiments, may reveal interesting phenomena, and allow harder testing of IR methods.

References

[1]
J. Allan, J. Callan, B. Croft, L. Ballesteros, J. Broglio, J. Xu & H. Shu. INQUERY at TREC 5. In E.M. Voorhees & D.K. Harrnan (Eds.), Information technology: The Fifth Text Retrieval Conference (TREC-5). Gaithersburg, MD: National Institute of Standards and Technology, 119-132, 1997.
[2]
D.C. Blair, & M.E. Maron. An evaluation of retrieval effecuveness for a full-text document-retrieval system. Communications of the A CM, 28(3): 289-299, 1985.
[3]
P. Borlund & P. Ingwersen. Measures of relative relevance and ranked half-life: Performance indicators for interactive IR. In W.B. Croft, A. Moffat, C.J. van Rijsbergen, R. Wilkinson & J. Zobel (Eds.), Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. New York: ACM, 324-.-331, 1998.
[4]
W.J. Conover. Practical nonparametric statistics (2nd ed.). New York: John Wiley & Sons, 1980.
[5]
R. Green. The expression of conceptual syntagmatic relationships: A comparative survey. Journal of Documentation, 51(4): 315-338, 1995.
[6]
W.R. Hersh & D.H. Hickam. An evaluation of interactive Boolean and natural language searching with an online medical textbook. Journal of the American Society for Information Science, 46(7): 478.--489, 1995.
[7]
P. Ingwersen & P. Willett. An introduction to algorithmic and cognitive approaches for information retrieval. Libri, 450: 160-177, 1995.
[8]
E.M. Keen. The use of term position devices in ranked output experiments. Journal of Documentation, 47(1): 1-22, 1991.
[9]
J. Kekiliinen. The effects of query complexity, expansion and structure on retrieval performance in probabilistic text retrieval. Ph.D. dissertation. Department of Information Studies, University of Tampere, 1999.
[10]
J. Kekalainen & K. Jarvelin. The co-effects of query structure and expansion on retrieval performance in probabilistic text retrieval. Information Retrieval, 1(4): 329-344, 2000.
[11]
J. Keklinen & K. Jrvelin. The impact of query structure and query expansion on retrieval performance. In W.B. Croft, A. Moffat, C.J. van Rijsbergen, R. Wilkinson & J. Zobei (Eds.), Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. New York: ACM, 130-137, 1998.
[12]
R.M. Losee. Text retrieval and filtering: Analytic models of performance. Kluwer Acadermc Publishers: Boston, 1998.
[13]
T.B. Rajashekar & W.B. Croft. Combining automatic and manual index representatzons m probabilistc retrieval. Journal of the American Society for Information Science, 46(4): 272-283, 1995.
[14]
S.E. Robertson & N.J. Belkin. Ranking in principle. Journal of Documentation, 34(2): 93-100, 1978.
[15]
T. Saracevic, P. Kantor, A. Chamis & D. Tnvison. A study of information seeking and retrieving. I. Background and methodology. Journal of the American Society for Information Science, 39(3): 161-176, 1988.
[16]
S. Smithson. Information retrieval evaluation in practice: A case study approach. Information Processing & Management, 30(2): 205-221, 1994.
[17]
E. Sormunen. A Method for Measuring Wtde Range PerformancIe of Boolean Queries in Full-Text Databases. Ph.D. dssertation. Department of Information Studies, University of Tampere, 2000.
[18]
H.R. Turtle. Inference networks for document retrieval. Ph.D. dissertation. Computer and information Science Department, University of Massachusetts, 1990.

Cited By

View all
  • (2024)An Architecture as an Alternative to Gradient Boosted Decision Trees for Multiple Machine Learning TasksElectronics10.3390/electronics1312229113:12(2291)Online publication date: 12-Jun-2024
  • (2024)GastroBot: a Chinese gastrointestinal disease chatbot based on the retrieval-augmented generationFrontiers in Medicine10.3389/fmed.2024.139255511Online publication date: 22-May-2024
  • (2024)A Blueprint of IR Evaluation Integrating Task and User CharacteristicsACM Transactions on Information Systems10.1145/367516242:6(1-38)Online publication date: 1-Jul-2024
  • Show More Cited By

Index Terms

  1. IR evaluation methods for retrieving highly relevant documents

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    SIGIR '00: Proceedings of the 23rd annual international ACM SIGIR conference on Research and development in information retrieval
    July 2000
    396 pages
    ISBN:1581132263
    DOI:10.1145/345508
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 01 July 2000

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Article

    Conference

    SIGIR00
    Sponsor:
    • Greek Com Soc
    • SIGIR
    • Athens U of Econ & Business

    Acceptance Rates

    Overall Acceptance Rate 792 of 3,983 submissions, 20%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)531
    • Downloads (Last 6 weeks)75
    Reflects downloads up to 14 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)An Architecture as an Alternative to Gradient Boosted Decision Trees for Multiple Machine Learning TasksElectronics10.3390/electronics1312229113:12(2291)Online publication date: 12-Jun-2024
    • (2024)GastroBot: a Chinese gastrointestinal disease chatbot based on the retrieval-augmented generationFrontiers in Medicine10.3389/fmed.2024.139255511Online publication date: 22-May-2024
    • (2024)A Blueprint of IR Evaluation Integrating Task and User CharacteristicsACM Transactions on Information Systems10.1145/367516242:6(1-38)Online publication date: 1-Jul-2024
    • (2024)The Impacts of Data, Ordering, and Intrinsic Dimensionality on Recall in Hierarchical Navigable Small WorldsProceedings of the 2024 ACM SIGIR International Conference on Theory of Information Retrieval10.1145/3664190.3672512(25-33)Online publication date: 2-Aug-2024
    • (2024)A Survey of Graph Neural Networks for Social Recommender SystemsACM Computing Surveys10.1145/366182156:10(1-34)Online publication date: 22-Jun-2024
    • (2024)Mitigating the Impact of Inaccurate Feedback in Dynamic Learning-to-Rank: A Study of Overlooked Interesting ItemsACM Transactions on Intelligent Systems and Technology10.1145/3653983Online publication date: 26-Mar-2024
    • (2024)Exploring Coresets for Efficient Training and Consistent Evaluation of Recommender SystemsProceedings of the 18th ACM Conference on Recommender Systems10.1145/3640457.3691716(1152-1157)Online publication date: 8-Oct-2024
    • (2024)Evaluating Retrieval Quality in Retrieval-Augmented GenerationProceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3626772.3657957(2395-2400)Online publication date: 10-Jul-2024
    • (2024)Deep Ranking Distribution Preserving Hashing for Robust Multi-Label Cross-Modal RetrievalIEEE Transactions on Multimedia10.1109/TMM.2024.335899526(7027-7042)Online publication date: 2024
    • (2024)Multisource Soft Labeling and Hard Negative Sampling for Retrieval Distractor RankingIEEE Transactions on Learning Technologies10.1109/TLT.2023.332554917(664-676)Online publication date: 1-Jan-2024
    • Show More Cited By

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Get Access

    Login options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media