Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/1015330.1015350acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlConference Proceedingsconference-collections
Article

Co-EM support vector learning

Published: 04 July 2004 Publication History

Abstract

Multi-view algorithms, such as co-training and co-EM, utilize unlabeled data when the available attributes can be split into independent and compatible subsets. Co-EM outperforms co-training for many problems, but it requires the underlying learner to estimate class probabilities, and to learn from probabilistically labeled data. Therefore, co-EM has so far only been studied with naive Bayesian learners. We cast linear classifiers into a probabilistic framework and develop a co-EM version of the Support Vector Machine. We conduct experiments on text classification problems and compare the family of semi-supervised support vector algorithms under different conditions, including violations of the assumptions underlying multi-view learning. For some problems, such as course web page classification, we observe the most accurate results reported so far.

References

[1]
Baluja, S. (1998). Probabilistic modeling for face orientation discrimination: Learning from labeled and unlabeled data. Advances in Neural Information Processing Systems.
[2]
Bennett, K. (1999). Combining support vector and mathematical programming methods for classification. Advances in Kernel Methods - Support Vector Learning. MIT Press.
[3]
Blum, A., & Mitchell, T. (1998). Combining labeled and unlabeled data with co-training. Proceedings of the Conference on Computational Learning Theory (pp. 92--100).
[4]
Bradley, A. (1997). The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognition, 30, 1145--1159.
[5]
Brefeld, U., Geibel, P., & Wysotzki, F. (2003). Support vector machines with example dependent costs. Proceedings of the European Conference on Machine Learning.
[6]
Collins, M., & Singer, Y. (1999). Unsupervised models for named entity classification. Proceedings of the Conference on Empirical Methods in Natural Language Processing.
[7]
Cooper, D., & Freeman, J. (1970). On the asymptotic improvement in the outcome of supervised learning provided by additional nonsupervised learning. IEEE Transactions on Computers, C-19, 1055--1063.
[8]
Cozman, F., Cohen, I., & Cirelo, M. (2003). Semi-supervised learning of mixture models. Proceedings of the International Conference on Machine Learning (pp. 99--106).
[9]
Dempster, A., Laird, N., & Rubin, D. (1977). Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society B, 39.
[10]
Denis, F., Laurent, A., Gilleron, R., & Tommasi, M. (2003). Text classification and co-training from positive and unlabeled examples. ICML Workshop on the Continuum from Labeled to Unlabeled Data.
[11]
Ghani, R. (2002). Combining labeled and unlabeled data for multiclass text categorization. Proceedings of the International Conference on Machine Learning.
[12]
Joachims, T. (1999a). Making large-scale SVM learning practical. Advances in Kernel Methods - Support Vector Learning. MIT Press.
[13]
Joachims, T. (1999b). Transductive inference for text classification using support vector machines. Proceedings of the International Conference on Machine Learning.
[14]
Joachims, T. (2003). Transductive learning via spectral graph partitioning. Proceedings of the International Conference on Machine Learning.
[15]
Kiritchenko, S., & Matwin, S. (2002). Email classification with co-training (Technical Report). University of Ottawa.
[16]
Kockelkorn, M., Lüneburg, A., & Scheffer, T. (2003). Using transduction and multi-view learning to answer emails. Proceedings of the European Conference on Principle and Practice of Knowledge Discovery in Databases.
[17]
McCallum, A., & Nigam, K. (1998). Employing EM in pool-based active learning for text classification. Proceedings of the International Conference on Machine Learning.
[18]
Mladenic, D. (2002). Learning word normalization using word suffix and context from unlabeled data. Proceedings of the International Conference on Machine Learning (pp. 427--434).
[19]
Muslea, I., Kloblock, C., & Minton, S. (2002a). Active + semi-supervised learning = robust multi-view learning. Proceedings of the International Conference on Machine Learning (pp. 435--442).
[20]
Muslea, I., Kloblock, C., & Minton, S. (2002b). Adaptive view validation: A first step towards automatic view detection. Proceedings of the International Conference on Machine Learning (pp. 443--450).
[21]
Nigam, K., & Ghani, R. (2000). Analyzing the effectiveness and applicability of co-training. Proceedings of Information and Knowledge Management.
[22]
Nigam, K., McCallum, A. K., Thrun, S., & Mitchell, T. M. (2000). Text classification from labeled and unlabeled documents using EM. Machine Learning, 39.
[23]
Provost, F., Fawcett, T., & Kohavi, R. (1998). The case against accuracy estimation for comparing inductive algorithms. Proceedings of the International Conference on Machine Learning (pp. 445--453).
[24]
Seeger, M. (2001). Learning with labeled and unlabeled data. (Technical Report). University of Edinburgh.
[25]
Shahshahani, B., & Landgrebe, D. (1994). The effect of unlabeled samples in reducing the small sample size problem and mitigating the Hughes phenomenon. IEEE Transactions on Geoscience and Remote Sensing, 32, 1087--1095.
[26]
Vapnik, V. (1998). Statistical learning theory. Wiley.

Cited By

View all
  • (2024)L1-penalized AUC-optimization with a surrogate lossCommunications for Statistical Applications and Methods10.29220/CSAM.2024.31.2.20331:2(203-212)Online publication date: 31-Mar-2024
  • (2024)CFARnet: Deep learning for target detection with constant false alarm rateSignal Processing10.1016/j.sigpro.2024.109543223(109543)Online publication date: Oct-2024
  • (2024)Appformer: A novel framework for mobile app usage prediction leveraging progressive multi-modal data fusion and feature extractionExpert Systems with Applications10.1016/j.eswa.2024.125903(125903)Online publication date: Nov-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICML '04: Proceedings of the twenty-first international conference on Machine learning
July 2004
934 pages
ISBN:1581138385
DOI:10.1145/1015330
  • Conference Chair:
  • Carla Brodley

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 04 July 2004

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Article

Acceptance Rates

Overall Acceptance Rate 140 of 548 submissions, 26%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)13
  • Downloads (Last 6 weeks)1
Reflects downloads up to 30 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2024)L1-penalized AUC-optimization with a surrogate lossCommunications for Statistical Applications and Methods10.29220/CSAM.2024.31.2.20331:2(203-212)Online publication date: 31-Mar-2024
  • (2024)CFARnet: Deep learning for target detection with constant false alarm rateSignal Processing10.1016/j.sigpro.2024.109543223(109543)Online publication date: Oct-2024
  • (2024)Appformer: A novel framework for mobile app usage prediction leveraging progressive multi-modal data fusion and feature extractionExpert Systems with Applications10.1016/j.eswa.2024.125903(125903)Online publication date: Nov-2024
  • (2024)Adversarial de-overlapping learning machines for supervised and semi-supervised learningInternational Journal of Machine Learning and Cybernetics10.1007/s13042-024-02389-9Online publication date: 7-Oct-2024
  • (2024)BSRU: boosting semi-supervised regressor through ramp-up unsupervised lossKnowledge and Information Systems10.1007/s10115-023-02044-966:5(2769-2797)Online publication date: 18-Jan-2024
  • (2024)SiamALNet: A Semi-supervised Siamese Neural Network with Active Learning Approach for Auto-LabelingProceedings of World Conference on Artificial Intelligence: Advances and Applications10.1007/978-981-97-4496-1_20(257-269)Online publication date: 1-Oct-2024
  • (2023)Factored Particle Swarm Optimization for Policy Co-training in Reinforcement LearningProceedings of the Genetic and Evolutionary Computation Conference10.1145/3583131.3590376(30-38)Online publication date: 15-Jul-2023
  • (2023)Self-Paced Co-Training of Graph Neural Networks for Semi-Supervised Node ClassificationIEEE Transactions on Neural Networks and Learning Systems10.1109/TNNLS.2022.315768834:11(9234-9247)Online publication date: Nov-2023
  • (2023)Multiview Latent Structure LearningKnowledge-Based Systems10.1016/j.knosys.2023.110707276:COnline publication date: 27-Sep-2023
  • (2023)Robust AUC maximization for classification with pairwise confidence comparisonsFrontiers of Computer Science10.1007/s11704-023-2709-518:4Online publication date: 16-Dec-2023
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media