• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations | Disambiguate

DMCA

Very simple classification rules perform well on most commonly used datasets (1993)

Cached

  • Download as a PDF

Download Links

  • [www.icar.cnr.it]
  • [staff.icar.cnr.it]
  • [dns2.icar.cnr.it]
  • [www.cs.uwyo.edu]
  • [www.cs.sunyit.edu]
  • [webdocs.cs.ualberta.ca]
  • [webdocs.cs.ualberta.ca]
  • [www.cs.ualberta.ca]
  • [sci2s.ugr.es]
  • [www.csi.uottawa.ca]

  • Other Repositories/Bibliography

  • DBLP
  • Save to List
  • Add to Collection
  • Correct Errors
  • Monitor Changes
by Robert C. Holte
Venue:Machine Learning
Citations:547 - 5 self
  • Summary
  • Citations
  • Active Bibliography
  • Co-citation
  • Clustered Documents
  • Version History

BibTeX

@INPROCEEDINGS{Holte93verysimple,
    author = {Robert C. Holte},
    title = {Very simple classification rules perform well on most commonly used datasets},
    booktitle = {Machine Learning},
    year = {1993},
    pages = {63--91}
}

Share

Facebook Twitter Reddit Bibsonomy

OpenURL

 

Abstract

The classification rules induced by machine learning systems are judged by two criteria: their classification accuracy on an independent test set (henceforth "accuracy"), and their complexity. The relationship between these two criteria is, of course, of keen interest to the machine learning community. There are in the literature some indications that very simple rules may achieve surprisingly high accuracy on many datasets. For example, Rendell occasionally remarks that many real world datasets have "few peaks (often just one) " and so are "easy to learn" (Rendell & Seshu, 1990, p.256). Similarly, Shavlik et al. (1991) report that, with certain qualifications, "the accuracy of the perceptron is hardly distinguishable from the more complicated learning algorithms " (p.134). Further evidence is provided by studies of pruning methods (e.g. Buntine & Niblett, 1992; Clark & Niblett, 1989; Mingers, 1989), where accuracy is rarely seen to decrease as pruning becomes more severe (for example, see Table 1) 1. This is so even when rules are pruned to the extreme, as happened with the "Err-comp " pruning method in Mingers (1989). This method produced the most accurate decision trees, and in four of the five domains studied these trees had only 2 or 3 leaves

Keyphrases

simple classification rule    classification rule    err-comp quot    complicated learning algorithm quot    rendell seshu    accurate decision tree    classification accuracy    clark niblett    accuracy quot    many datasets    henceforth quot    simple rule    buntine niblett    high accuracy    keen interest    many real world datasets    certain qualification    independent test set   

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University