Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/2983323.2983813acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

Compression-Based Selective Sampling for Learning to Rank

Published: 24 October 2016 Publication History

Abstract

Learning to rank (L2R) algorithms use a labeled training set to generate a ranking model that can be later used to rank new query results. These training sets are very costly and laborious to produce, requiring human annotators to assess the relevance or order of the documents in relation to a query. Active learning (AL) algorithms are able to reduce the labeling effort by actively sampling an unlabeled set and choosing data instances that maximize the effectiveness of a learning function. But AL methods require constant supervision, as documents have to be labeled at each round of the process. In this paper, we propose that certain characteristics of unlabeled L2R datasets allow for an unsupervised, compression-based selection process to be used to create small and yet highly informative and effective initial sets that can later be labeled and used to bootstrap a L2R system. We implement our ideas through a novel unsupervised selective sampling method, which we call Cover, that has several advantages over AL methods tailored to L2R. First, it does not need an initial labeled seed set and can select documents from scratch. Second, selected documents do not need to be labeled as the iterations of the method progress since it is unsupervised (i.e., no learning model needs to be updated). Thus, an arbitrarily sized training set can be selected without human intervention depending on the available budget. Third, the method is efficient and can be run on unlabeled collections containing millions of query-document instances. We run various experiments with two important L2R benchmarking collections to show that the proposed method allows for the creation of small, yet very effective training sets. It achieves full training-like performance with less than 10% of the original sets selected, outperforming the baselines in both effectiveness and scalability.

References

[1]
G. Babu and E. Feigelson. Astrostatistics. Chapman & Hall/CRC Interdisciplinary Statistics. Taylor & Francis, 1996.
[2]
K. Brinker. Incorporating diversity in active learning with support vector machines. ICML '03, pages 59--66, 2003.
[3]
P. Cai, W. Gao, A. Zhou, and K. Wong. Relevant knowledge helps in choosing right teacher. SIGIR '11, pages 115--124, 2011.
[4]
W. Cai and Y. Zhang. Variance maximization via noise injection for active sampling in learning to rank. CIKM '12, pages 1809--1813, 2012.
[5]
Y. Cheng, Z. Chen, L. Liu, J. Wang, A. Agrawal, and A. Choudhary. Feedback-driven multiclass active learning for data streams. CIKM '13, page 1311--1320, 2013.
[6]
T. M. Cover and J. A. Thomas. Elements of Information Theory (Wiley Series in Telecommunications and Signal Processing). Wiley-Interscience, 2006.
[7]
P. Donmez and J. G. Carbonell. Optimizing estimated loss reduction for active sampling in rank learning. ICML '08, pages 248--255, 2008.
[8]
P. Donmez and J. G. Carbonell. Active sampling for rank learning via optimizing the area under the ROC curve. ECIR '09, pages 78--89, 2009.
[9]
X. Geng, T.-Y. Liu, T. Qin, and H. Li. Feature Selection for Ranking. SIGIR '07, pages 407--414, 2007.
[10]
X. Geng, T. Qin, T. Liu, X. Cheng, and H. Li. Selecting optimal training data for learning to rank. Inf. Process. and Manage., 47(5):730--741, Sept. 2011.
[11]
B. Long, O. Chapelle, Y. Zhang, Y. Chang, Z. Zheng, and B. Tseng. Active learning for ranking through expected loss optimization. SIGIR '10, pages 267--274, 2010.
[12]
D. J. C. MacKay. Information Theory, Inference & Learning Algorithms. Cambridge University Press, New York, NY, USA, 2002.
[13]
W. B. March, P. Ram, and A. G. Gray. Fast euclidean minimum spanning tree: Algorithm, analysis, and applications. KDD '10, pages 603--612, 2010.
[14]
R. Mehrotra and E. Yilmaz. Representative & Informative Query Selection for Learning to Rank Using Submodular Functions. SIGIR '15, pages 545--554, 2015.
[15]
A. Mohan, Z. Chen, and K. Q. Weinberger. Web-Search Ranking with Initialized Gradient Boosted Regression Trees. Journal of Machine Learning Research, 14:77--89, 2011.
[16]
F. Pan, T. Converse, D. Ahn, F. Salvetti, and G. Donato. Feature Selection for Ranking Using Boosted Trees. CIKM '09, pages 2025--2028, 2009.
[17]
T. Qin, T. Liu, J. Xu, and H. Li. LETOR: a benchmark collection for research on learning to rank for information retrieval. Information Retrieval, 13:346--374, Aug. 2010.
[18]
F. Radlinski and T. Joachims. Active exploration for learning rankings from clickthrough data. SIGKDD '07, pages 570--579, 2007.
[19]
C. E. Shannon. A mathematical theory of communication. Bell System Technical Journal, 27:379--423, 623--656, 1948.
[20]
R. Silva, M. A. Gonçalves, and A. Veloso. Rule-based active sampling for learning to rank. ECML PKDD '11, pages 240--255, 2011.
[21]
R. M. Silva, M. A. Gonçalves, and A. Veloso. A two-stage active learning method for learning to rank. JASIST, 65(1):109--128, Jan. 2014.
[22]
Z. Xu, R. Akella, and Y. Zhang. Incorporating Diversity and Density in Active Learning for Relevance Feedback. ECIR'07, pages 246--257, 2007.
[23]
L. Yang, L. Wang, B. Geng, and X.-S. Hua. Query Sampling for Ranking Learning in Web Search. SIGIR '09, pages 754--755, 2009.
[24]
E. Yilmaz and S. Robertson. Deep Versus Shallow Judgments in Learning to Rank. SIGIR '09, pages 662--663, 2009.
[25]
H. Yu. SVM selective sampling for ranking with application to data retrieval. SIGKDD '05, pages 354--363, 2005.

Cited By

View all
  • (2023)Reducing the user labeling effort in effective high recall tasks by fine-tuning active learningJournal of Intelligent Information Systems10.1007/s10844-022-00772-y61:2(453-472)Online publication date: 19-Jan-2023
  • (2022)Looking Back on the Past: Active Learning With Historical Evaluation ResultsIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2020.304581634:10(4921-4932)Online publication date: 1-Oct-2022
  • (2019)Adaptive geometric median prototype selection method for k-nearest neighbors classificationIntelligent Data Analysis10.3233/IDA-18419023:4(855-876)Online publication date: 26-Sep-2019
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
CIKM '16: Proceedings of the 25th ACM International on Conference on Information and Knowledge Management
October 2016
2566 pages
ISBN:9781450340731
DOI:10.1145/2983323
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 24 October 2016

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. active learning
  2. compression
  3. learning to rank
  4. selective sampling

Qualifiers

  • Research-article

Funding Sources

  • InWeb
  • CNPq
  • MASWeb

Conference

CIKM'16
Sponsor:
CIKM'16: ACM Conference on Information and Knowledge Management
October 24 - 28, 2016
Indiana, Indianapolis, USA

Acceptance Rates

CIKM '16 Paper Acceptance Rate 160 of 701 submissions, 23%;
Overall Acceptance Rate 1,861 of 8,427 submissions, 22%

Upcoming Conference

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)4
  • Downloads (Last 6 weeks)0
Reflects downloads up to 19 Sep 2024

Other Metrics

Citations

Cited By

View all
  • (2023)Reducing the user labeling effort in effective high recall tasks by fine-tuning active learningJournal of Intelligent Information Systems10.1007/s10844-022-00772-y61:2(453-472)Online publication date: 19-Jan-2023
  • (2022)Looking Back on the Past: Active Learning With Historical Evaluation ResultsIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2020.304581634:10(4921-4932)Online publication date: 1-Oct-2022
  • (2019)Adaptive geometric median prototype selection method for k-nearest neighbors classificationIntelligent Data Analysis10.3233/IDA-18419023:4(855-876)Online publication date: 26-Sep-2019
  • (2017)LEARning Next gEneration Rankers (LEARNER 2017)Proceedings of the ACM SIGIR International Conference on Theory of Information Retrieval10.1145/3121050.3121110(331-332)Online publication date: 1-Oct-2017
  • (2017)The Impact of Linkage Methods in Hierarchical Clustering for Active Learning to RankProceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3077136.3080684(941-944)Online publication date: 7-Aug-2017

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media