Nothing Special   »   [go: up one dir, main page]

Skip to main content
Log in

Accurately learning from few examples with a polyhedral classifier

  • Published:
Computational Optimization and Applications Aims and scope Submit manuscript

Abstract

In the context of learning theory many efforts have been devoted to developing classification algorithms able to scale up with massive data problems. In this paper the complementary issue is addressed, aimed at deriving powerful classification rules by accurately learning from few data. This task is accomplished by solving a new mixed integer programming model that extends the notion of discrete support vector machines, in order to derive an optimal set of separating hyperplanes for binary classification problems. According to the cardinality of the set of hyperplanes, the classification region may take the form of a convex polyhedron or a polytope in the original space where the examples are defined. Computational tests on benchmark datasets highlight the effectiveness of the proposed model, that yields the greatest accuracy when compared to other classification approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Anthony, M.: On data classification by iterative linear partitioning. Discrete Appl. Math. 144, 2–16 (2004)

    Article  MATH  MathSciNet  Google Scholar 

  2. Astorino, A., Gaudioso, M.: Polyhedral separability through successive LP. J. Optim. Theory Appl. 112, 265–293 (2002)

    Article  MATH  MathSciNet  Google Scholar 

  3. Berg, C., Christensen, J.P.R., Ressel, P.: Harmonic Analysis on Semigroups: Theory of Positive Definite and Related Functions. Springer, New York (1984)

    MATH  Google Scholar 

  4. Bradley, P.S., Fayyad, U.M., Mangasarian, O.L.: Mathematical programming for data mining: formulations and challenges. INFORMS J. Comput. 11, 217–238 (1999)

    Article  MATH  MathSciNet  Google Scholar 

  5. Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines (2001)

  6. Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods. Cambridge University Press, Cambridge (2000)

    Google Scholar 

  7. Evgeniou, T., Poggio, T., Pontil, M., Verri, A.: Regularization and statistical learning theory for data analysis. J. Comput. Stat. Data Anal. 38, 421–432 (2002)

    Article  MATH  MathSciNet  Google Scholar 

  8. Geman, S.: Compositionality. Technical report, Brown University Faculty Bulletin (1999)

  9. Herman, G.T., Yeung, K.T.D.: On piecewise-linear classification. IEEE Trans. Pattern Anal. Mach. Intell. 14, 782–786 (1992)

    Article  Google Scholar 

  10. Hettich, S., Blake, C., Merz, C.: UCI repository of machine learning databases. URL http://www.ics.uci.edu/~mlearn/MLRepository.html (1998)

  11. Kohavi, R.: A study of cross-validation and bootstrap for accuracy estimation and model selection. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence, pp. 1137–1143. Morgan Kaufmann, San Mateo (1995)

    Google Scholar 

  12. Mangasarian, O.: Mathematical programming in data mining. Data Min. Knowl. Discov. 1, 183–201 (1997)

    Article  Google Scholar 

  13. Orsenigo, C., Vercellis, C.: Multivariate classification trees based on minimum features discrete support vector machines. IMA J. Manag. Math. 14, 221–234 (2003)

    Article  MATH  MathSciNet  Google Scholar 

  14. Orsenigo, C., Vercellis, C.: Discrete support vector decision trees via tabu-search. J. Comput. Stat. Data Anal. 47, 311–322 (2004)

    Article  MathSciNet  MATH  Google Scholar 

  15. Orsenigo, C., Vercellis, C.: One-against-all multicategory classification via discrete support vector machines. In: Ebecken N., et al. (eds.) Data Mining IV, pp. 255–264. WIT, Ashurst Lodge (2004)

    Google Scholar 

  16. Remshagen, A., Truemper, K.: Algorithms for logic-based abduction. Technical report, University of Texas at Dallas (2002)

  17. Scholkopf, B., Smola, A.J.: Learning with Kernels: Support Vector Machines, Regularization, Optimization, and beyond. MIT, Cambridge (2002)

    Google Scholar 

  18. Tikhonov, A.N., Arsenin, V.Y.: Solutions of Ill-Posed Problems. Winston, Washington (1977)

    MATH  Google Scholar 

  19. Vapnik, V.: The Nature of Statistical Learning Theory. Springer, New York (1995)

    MATH  Google Scholar 

  20. Vapnik, V.: Statistical Learning Theory. Wiley, New York (1998)

    MATH  Google Scholar 

  21. Wolberg, W.H., Mangasarian, O.L.: Multisurface method of pattern separation for medical diagnosis applied to breast cytology. Proc. Nat. Acad. Sci. 87, 9193–9196 (1990)

    Article  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Carlo Vercellis.

Additional information

This research was partially supported by PRIN grant 2004132117.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Orsenigo, C., Vercellis, C. Accurately learning from few examples with a polyhedral classifier. Comput Optim Appl 38, 235–247 (2007). https://doi.org/10.1007/s10589-007-9041-0

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10589-007-9041-0

Keywords

Navigation