• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 744,553
Next 10 →

Attribute Relevance in Multiclass Data Sets Using the Naive Bayes Rule∗

by J. M. Sotoca, J. S. Sánchez, F. Pla
"... Feature selection using the naive Bayes rule is presented for the case of multiclass data sets. In this paper, the EM algorithm is applied to each class projected over the fea-tures in order to obtain an estimation of the class proba-bility density function. A matrix of weights per class and feature ..."
Abstract - Add to MetaCart
Feature selection using the naive Bayes rule is presented for the case of multiclass data sets. In this paper, the EM algorithm is applied to each class projected over the fea-tures in order to obtain an estimation of the class proba-bility density function. A matrix of weights per class

Naive (Bayes) at Forty: The Independence Assumption in Information Retrieval

by David D. Lewis , 1998
"... The naive Bayes classifier, currently experiencing a renaissance in machine learning, has long been a core technique in information retrieval. We review some of the variations of naive Bayes models used for text retrieval and classification, focusing on the distributional assump- tions made abou ..."
Abstract - Cited by 496 (1 self) - Add to MetaCart
The naive Bayes classifier, currently experiencing a renaissance in machine learning, has long been a core technique in information retrieval. We review some of the variations of naive Bayes models used for text retrieval and classification, focusing on the distributional assump- tions made

On Discriminative vs. Generative classifiers: A comparison of logistic regression and naive Bayes

by Andrew Y. Ng, Michael I. Jordan , 2001
"... We compare discriminative and generative learning as typified by logistic regression and naive Bayes. We show, contrary to a widely held belief that discriminative classifiers are almost always to be preferred, that there can often be two distinct regimes of performance as the training set size is i ..."
Abstract - Cited by 513 (8 self) - Add to MetaCart
We compare discriminative and generative learning as typified by logistic regression and naive Bayes. We show, contrary to a widely held belief that discriminative classifiers are almost always to be preferred, that there can often be two distinct regimes of performance as the training set size

A comparison of event models for Naive Bayes text classification

by Andrew McCallum, Kamal Nigam , 1998
"... Recent work in text classification has used two different first-order probabilistic models for classification, both of which make the naive Bayes assumption. Some use a multi-variate Bernoulli model, that is, a Bayesian Network with no dependencies between words and binary word features (e.g. Larkey ..."
Abstract - Cited by 1002 (27 self) - Add to MetaCart
Recent work in text classification has used two different first-order probabilistic models for classification, both of which make the naive Bayes assumption. Some use a multi-variate Bernoulli model, that is, a Bayesian Network with no dependencies between words and binary word features (e

Bayes Factors

by Robert E. Kass, Adrian E. Raftery , 1995
"... In a 1935 paper, and in his book Theory of Probability, Jeffreys developed a methodology for quantifying the evidence in favor of a scientific theory. The centerpiece was a number, now called the Bayes factor, which is the posterior odds of the null hypothesis when the prior probability on the null ..."
Abstract - Cited by 1766 (74 self) - Add to MetaCart
is one-half. Although there has been much discussion of Bayesian hypothesis testing in the context of criticism of P -values, less attention has been given to the Bayes factor as a practical tool of applied statistics. In this paper we review and discuss the uses of Bayes factors in the context of five

Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers

by Erin L. Allwein, Robert E. Schapire, Yoram Singer - JOURNAL OF MACHINE LEARNING RESEARCH , 2000
"... We present a unifying framework for studying the solution of multiclass categorization problems by reducing them to multiple binary problems that are then solved using a margin-based binary learning algorithm. The proposed framework unifies some of the most popular approaches in which each class ..."
Abstract - Cited by 560 (20 self) - Add to MetaCart
We present a unifying framework for studying the solution of multiclass categorization problems by reducing them to multiple binary problems that are then solved using a margin-based binary learning algorithm. The proposed framework unifies some of the most popular approaches in which each class

Feature selection based on mutual information: Criteria of max-depe ndency, max-relevance, and min-redundancy

by Hanchuan Peng, Fuhui Long, Chris Ding - IEEE Trans. Pattern Analysis and Machine Intelligence
"... Abstract—Feature selection is an important problem for pattern classification systems. We study how to select good features according to the maximal statistical dependency criterion based on mutual information. Because of the difficulty in directly implementing the maximal dependency condition, we f ..."
Abstract - Cited by 533 (7 self) - Add to MetaCart
to select a compact set of superior features at very low cost. We perform extensive experimental comparison of our algorithm and other methods using three different classifiers (naive Bayes, support vector machine, and linear discriminate analysis) and four different data sets (handwritten digits

Estimating Attributes: Analysis and Extensions of RELIEF

by Igor Kononenko , 1994
"... . In the context of machine learning from examples this paper deals with the problem of estimating the quality of attributes with and without dependencies among them. Kira and Rendell (1992a,b) developed an algorithm called RELIEF, which was shown to be very efficient in estimating attributes. Origi ..."
Abstract - Cited by 450 (23 self) - Add to MetaCart
. Original RELIEF can deal with discrete and continuous attributes and is limited to only two-class problems. In this paper RELIEF is analysed and extended to deal with noisy, incomplete, and multi-class data sets. The extensions are verified on various artificial and one well known real-world problem. 1

Dynamic Itemset Counting and Implication Rules for Market Basket Data

by Sergey Brin, Rajeev Motwani, Jeffrey D. Ullman, Shalom Tsur , 1997
"... We consider the problem of analyzing market-basket data and present several important contributions. First, we present a new algorithm for finding large itemsets which uses fewer passes over the data than classic algorithms, and yet uses fewer candidate itemsets than methods based on sampling. We in ..."
Abstract - Cited by 599 (6 self) - Add to MetaCart
We consider the problem of analyzing market-basket data and present several important contributions. First, we present a new algorithm for finding large itemsets which uses fewer passes over the data than classic algorithms, and yet uses fewer candidate itemsets than methods based on sampling. We

Sparse Bayesian Learning and the Relevance Vector Machine

by Michael E. Tipping, Alex Smola , 2001
"... This paper introduces a general Bayesian framework for obtaining sparse solutions to regression and classication tasks utilising models linear in the parameters. Although this framework is fully general, we illustrate our approach with a particular specialisation that we denote the `relevance vec ..."
Abstract - Cited by 958 (5 self) - Add to MetaCart
This paper introduces a general Bayesian framework for obtaining sparse solutions to regression and classication tasks utilising models linear in the parameters. Although this framework is fully general, we illustrate our approach with a particular specialisation that we denote the `relevance
Next 10 →
Results 1 - 10 of 744,553
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University