• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 2,784
Next 10 →

A Heteroskedasticity-Consistent Covariance Matrix Estimator And A Direct Test For Heteroskedasticity

by Halbert White , 1980
"... This paper presents a parameter covariance matrix estimator which is consistent even when the disturbances of a linear regression model are heteroskedastic. This estimator does not depend on a formal model of the structure of the heteroskedasticity. By comparing the elements of the new estimator ..."
Abstract - Cited by 3211 (5 self) - Add to MetaCart
to those of the usual covariance estimator, one obtains a direct test for heteroskedasticity, since in the absence of heteroskedasticity, the two estimators will be approximately equal, but will generally diverge otherwise. The test has an appealing least squares interpretation

Benchmarking Least Squares Support Vector Machine Classifiers

by Tony Van Gestel, Johan A. K. Suykens, Bart Baesens, Stijn Viaene, Jan Vanthienen, Guido Dedene, Bart De Moor, Joos Vandewalle - NEURAL PROCESSING LETTERS , 2001
"... In Support Vector Machines (SVMs), the solution of the classification problem is characterized by a (convex) quadratic programming (QP) problem. In a modified version of SVMs, called Least Squares SVM classifiers (LS-SVMs), a least squares cost function is proposed so as to obtain a linear set of eq ..."
Abstract - Cited by 476 (46 self) - Add to MetaCart
In Support Vector Machines (SVMs), the solution of the classification problem is characterized by a (convex) quadratic programming (QP) problem. In a modified version of SVMs, called Least Squares SVM classifiers (LS-SVMs), a least squares cost function is proposed so as to obtain a linear set

Regression Shrinkage and Selection Via the Lasso

by Robert Tibshirani - JOURNAL OF THE ROYAL STATISTICAL SOCIETY, SERIES B , 1994
"... We propose a new method for estimation in linear models. The "lasso" minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant. Because of the nature of this constraint it tends to produce some coefficients that are exactl ..."
Abstract - Cited by 4212 (49 self) - Add to MetaCart
We propose a new method for estimation in linear models. The "lasso" minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant. Because of the nature of this constraint it tends to produce some coefficients

Greedy Function Approximation: A Gradient Boosting Machine

by Jerome H. Friedman - Annals of Statistics , 2000
"... Function approximation is viewed from the perspective of numerical optimization in function space, rather than parameter space. A connection is made between stagewise additive expansions and steepest{descent minimization. A general gradient{descent \boosting" paradigm is developed for additi ..."
Abstract - Cited by 1000 (13 self) - Add to MetaCart
for additive expansions based on any tting criterion. Specic algorithms are presented for least{squares, least{absolute{deviation, and Huber{M loss functions for regression, and multi{class logistic likelihood for classication. Special enhancements are derived for the particular case where the individual

Algorithms for Non-negative Matrix Factorization

by Daniel D. Lee, H. Sebastian Seung - In NIPS , 2001
"... Non-negative matrix factorization (NMF) has previously been shown to be a useful decomposition for multivariate data. Two different multiplicative algorithms for NMF are analyzed. They differ only slightly in the multiplicative factor used in the update rules. One algorithm can be shown to minim ..."
Abstract - Cited by 1246 (5 self) - Add to MetaCart
to minimize the conventional least squares error while the other minimizes the generalized Kullback-Leibler divergence. The monotonic convergence of both algorithms can be proven using an auxiliary function analogous to that used for proving convergence of the ExpectationMaximization algorithm

Clustering with Bregman Divergences

by Arindam Banerjee, Srujana Merugu, Inderjit Dhillon, Joydeep Ghosh - JOURNAL OF MACHINE LEARNING RESEARCH , 2005
"... A wide variety of distortion functions are used for clustering, e.g., squared Euclidean distance, Mahalanobis distance and relative entropy. In this paper, we propose and analyze parametric hard and soft clustering algorithms based on a large class of distortion functions known as Bregman divergence ..."
Abstract - Cited by 443 (57 self) - Add to MetaCart
A wide variety of distortion functions are used for clustering, e.g., squared Euclidean distance, Mahalanobis distance and relative entropy. In this paper, we propose and analyze parametric hard and soft clustering algorithms based on a large class of distortion functions known as Bregman

A Least Squares Interpretation of Sub-Space Methods for System Identification.

by Lennart Ljung, Tomas McKelvey - In: Proc. IEEE Conference on Decision and Control, CDC.. Kobe , 1996
"... So called subspace methods for direct identification of linear models in state space form have drawn considerable interest recently. The algorithms consist of series of quite complex projections, and it is not so easy to intuitively understand how they work. They have also defied, so far, complete a ..."
Abstract - Cited by 4 (1 self) - Add to MetaCart
asymptotic analysis of their stochastic properties. This contribution describes an interpretation of how they work. It specifically deals how consistent estimates of the dynamics can be achieved, even though correct predictors are not used. We stress how the basic idea is to focus on the estimation

The Determinants of Credit Spread Changes.

by Pierre Collin-Dufresne , Robert S Goldstein , J Spencer Martin , Gurdip Bakshi , Greg Bauer , Dave Brown , Francesca Carrieri , Peter Christoffersen , Susan Christoffersen , Greg Duffee , Darrell Duffie , Vihang Errunza , Gifford Fong , Mike Gallmeyer , Laurent Gauthier , Rick Green , John Griffin , Jean Helwege , Kris Jacobs , Chris Jones , Andrew Karolyi , Dilip Madan , David Mauer , Erwan Morellec , Federico Nardari , N R Prabhala , Tony Sanders , Sergei Sarkissian , Bill Schwert , Ken Singleton , Chester Spatt , René Stulz - Journal of Finance , 2001
"... ABSTRACT Using dealer's quotes and transactions prices on straight industrial bonds, we investigate the determinants of credit spread changes. Variables that should in theory determine credit spread changes have rather limited explanatory power. Further, the residuals from this regression are ..."
Abstract - Cited by 422 (2 self) - Add to MetaCart
rates, r 10 t . To capture potential non-linear effects due to convexity, we also include the squared level of the term structure, (r 10 t ) 2 . Slope of Yield Curve We define the slope of the yield curve as the difference between Datastream's 10-year and 2-year Benchmark Treasury yields, slope

The log of Gravity

by Joao Santos Silva , Silvana Tenreyro - THE REVIEW OF ECONOMICS AND STATISTICS , 2005
"... Although economists have long been aware of Jensen's inequality, many econometric applications have neglected an important implication of it: the standard practice of interpreting the parameters of log-linearized models estimated by ordinary least squares as elasticities can be highly misleadin ..."
Abstract - Cited by 333 (6 self) - Add to MetaCart
Although economists have long been aware of Jensen's inequality, many econometric applications have neglected an important implication of it: the standard practice of interpreting the parameters of log-linearized models estimated by ordinary least squares as elasticities can be highly

Robust Solutions To Least-Squares Problems With Uncertain Data

by Laurent El Ghaoui, Hervé Lebret , 1997
"... . We consider least-squares problems where the coefficient matrices A; b are unknown-butbounded. We minimize the worst-case residual error using (convex) second-order cone programming, yielding an algorithm with complexity similar to one singular value decomposition of A. The method can be interpret ..."
Abstract - Cited by 205 (14 self) - Add to MetaCart
. We consider least-squares problems where the coefficient matrices A; b are unknown-butbounded. We minimize the worst-case residual error using (convex) second-order cone programming, yielding an algorithm with complexity similar to one singular value decomposition of A. The method can
Next 10 →
Results 1 - 10 of 2,784
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University