Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3132847.3133070acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
short-paper

Compact Multiple-Instance Learning

Published: 06 November 2017 Publication History

Abstract

The weakly supervised Multiple-Instance Learning (MIL) problem has been successfully applied in information retrieval tasks. Two related issues might affect the performance of MIL algorithms: how to cope with label ambiguities and how to deal with non-discriminative components, and we propose COmpact MultiPle-Instance LEarning (COMPILE) to consider them simultaneously. To treat label ambiguities, COMPILE seeks ground-truth positive instances in positive bags. By using weakly supervised information to learn data's short binary representations, COMPILE enhances discrimination via strengthening discriminative components and suppressing non-discriminative ones. We adapt block coordinate descent to optimize COMPILE efficiently. Experiments on text categorization empirically show: 1) COMPILE unifies disambiguation and data preprocessing successfully; 2) it generates short binary representations efficiently to enhance discrimination at significantly reduced storage cost.

References

[1]
J. Amores. 2013. Multiple instance classification: Review, taxonomy and comparative study. Artif. Intell. Vol. 201 (2013), 81--105.
[2]
S. Andrews, I. Tsochantaridis, and T. Hofmann. 2003. Support vector machines for multiple-instance learning NIPS. 577--584.
[3]
J.F. Bonnans, J.C. Gilbert, C. Lemaréchal, and C.A. Sagastizábal. 2006. Numerical optimization: theoretical and practical aspects. Springer Science & Business Media.
[4]
J. Chai, H. Chen, L. Huang, and F. Shang. 2014. Maximum margin multiple-instance feature weighting. Pattern Recog., Vol. 47, 6 (2014), 2091--2103.
[5]
R.O. Duda, P.E. Hart, and D.G. Stork. 2012. Pattern classification. John Wiley & Sons.
[6]
Y. Gong and S. Lazebnik. 2011. Iterative quantization: A procrustean approach to learning binary codes CVPR. 817--824.
[7]
Wu-Jun Li and Dit-Yan Yeung. 2010. MILD: Multiple-Instance Learning via Disambiguation. IEEE Trans. Knowl. Data Eng. Vol. 22, 1 (2010), 76--89.
[8]
F. Shen, C. Shen, W. Liu, and H.T. Shen. 2015. Supervised discrete hashing. In CVPR. 37--45.
[9]
X. Shen, W. Liu, I. Tsang, F. Shen, and Q.S. Sun. 2017. Compressed K-Means for Large-Scale Clustering. In AAAI. 2527--2533.
[10]
J. Wang and J.D. Zucker. 2000. Solving the Multiple-Instance Problem: A Lazy Learning Approach ICML. 1119--1126.
[11]
J. Wu, Z. Hong, S. Pan, X. Zhu, Z. Cai, and C. Zhang. 2014. Exploring features for complicated objects: Cross-view feature selection for multi-instance learning. In CIKM. 1699--1708.
[12]
M.L. Zhang and Z.H. Zhou. 2006. Adapting RBF neural networks to multi-instance learning. Neural Processing Letters Vol. 23, 1 (2006), 1--26.
[13]
Q. Zhang, S.A. Goldman, et almbox. 2001. EM-DD: An improved multiple-instance learning technique NIPS. 1073--1080.
[14]
Y. Zhou, Z. Jorgensen, and M. Inge. 2007. Combating good word attacks on statistical spam filters with multiple instance learning ICTAI. 298--305.
[15]
Z.H. Zhou, K. Jiang, and M. Li. 2005. Multi-instance learning based web mining. Appl. Intell., Vol. 22, 2 (2005), 135--147.
[16]
Z.H. Zhou, Y.Y. Sun, and Y.F. Li. 2009. Multi-instance learning by treating instances as non-iid samples ICML. 1249--1256.

Cited By

View all
  • (2022)Multiple Instance Learning for Uplift ModelingProceedings of the 31st ACM International Conference on Information & Knowledge Management10.1145/3511808.3557655(4727-4731)Online publication date: 17-Oct-2022

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
CIKM '17: Proceedings of the 2017 ACM on Conference on Information and Knowledge Management
November 2017
2604 pages
ISBN:9781450349185
DOI:10.1145/3132847
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 06 November 2017

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. disambiguation
  2. multiple-instance learning
  3. storage cost
  4. text categorization

Qualifiers

  • Short-paper

Funding Sources

  • The ARC Grant
  • The National Natural Science Foundation of China
  • The DP
  • The ARC Future Fellowship

Conference

CIKM '17
Sponsor:

Acceptance Rates

CIKM '17 Paper Acceptance Rate 171 of 855 submissions, 20%;
Overall Acceptance Rate 1,861 of 8,427 submissions, 22%

Upcoming Conference

CIKM '25

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)2
  • Downloads (Last 6 weeks)1
Reflects downloads up to 18 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2022)Multiple Instance Learning for Uplift ModelingProceedings of the 31st ACM International Conference on Information & Knowledge Management10.1145/3511808.3557655(4727-4731)Online publication date: 17-Oct-2022

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media