• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 22,725
Next 10 →

Exact Matrix Completion via Convex Optimization

by Emmanuel J. Candès, Benjamin Recht , 2008
"... We consider a problem of considerable practical interest: the recovery of a data matrix from a sampling of its entries. Suppose that we observe m entries selected uniformly at random from a matrix M. Can we complete the matrix and recover the entries that we have not seen? We show that one can perfe ..."
Abstract - Cited by 873 (26 self) - Add to MetaCart
We consider a problem of considerable practical interest: the recovery of a data matrix from a sampling of its entries. Suppose that we observe m entries selected uniformly at random from a matrix M. Can we complete the matrix and recover the entries that we have not seen? We show that one can

Algorithms for Non-negative Matrix Factorization

by Daniel D. Lee, H. Sebastian Seung - In NIPS , 2001
"... Non-negative matrix factorization (NMF) has previously been shown to be a useful decomposition for multivariate data. Two different multiplicative algorithms for NMF are analyzed. They differ only slightly in the multiplicative factor used in the update rules. One algorithm can be shown to minim ..."
Abstract - Cited by 1246 (5 self) - Add to MetaCart
Non-negative matrix factorization (NMF) has previously been shown to be a useful decomposition for multivariate data. Two different multiplicative algorithms for NMF are analyzed. They differ only slightly in the multiplicative factor used in the update rules. One algorithm can be shown

MATRIX FACTORIZATION TECHNIQUES FOR RECOMMENDER SYSTEMS

by Yehuda Koren, Robert Bell, Chris Volinsky - IEEE COMPUTER , 2009
"... As the Netflix Prize competition has demonstrated, matrix factorization models are superior to classic nearest-neighbor techniques for producing product recommendations, allowing the incorporation of additional information such as implicit feedback, temporal effects, and confidence levels. Modern co ..."
Abstract - Cited by 593 (4 self) - Add to MetaCart
As the Netflix Prize competition has demonstrated, matrix factorization models are superior to classic nearest-neighbor techniques for producing product recommendations, allowing the incorporation of additional information such as implicit feedback, temporal effects, and confidence levels. Modern

Learning the Kernel Matrix with Semi-Definite Programming

by Gert R. G. Lanckriet, Nello Cristianini, Laurent El Ghaoui, Peter Bartlett, Michael I. Jordan , 2002
"... Kernel-based learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information ..."
Abstract - Cited by 775 (21 self) - Add to MetaCart
problems in machine learning. In this paper we show how the kernel matrix can be learned from data via semi-definite programming (SDP) techniques. When applied

Non-negative matrix factorization with sparseness constraints,”

by Patrik O Hoyer , Patrik Hoyer@helsinki , Fi - Journal of Machine Learning Research, , 2004
"... Abstract Non-negative matrix factorization (NMF) is a recently developed technique for finding parts-based, linear representations of non-negative data. Although it has successfully been applied in several applications, it does not always result in parts-based representations. In this paper, we sho ..."
Abstract - Cited by 498 (0 self) - Add to MetaCart
Abstract Non-negative matrix factorization (NMF) is a recently developed technique for finding parts-based, linear representations of non-negative data. Although it has successfully been applied in several applications, it does not always result in parts-based representations. In this paper, we

DATA MATRIX

by unknown authors
"... Form Approved OMB No. 2070-0060 Paperwork Reduction Act Notice: The public reporting burden for this collection of information is estimated to average 0.25 hours per response for registration activities and 0.25 hours per response for reregistration and special review activities, including time for ..."
Abstract - Add to MetaCart
Form Approved OMB No. 2070-0060 Paperwork Reduction Act Notice: The public reporting burden for this collection of information is estimated to average 0.25 hours per response for registration activities and 0.25 hours per response for reregistration and special review activities, including time for reading the instructions and completing the necessary forms. Send comments regarding the burden estimate or any other aspect of this collection of

A Data Locality Optimizing Algorithm

by Michael E. Wolf, Monica S. Lam , 1991
"... This paper proposes an algorithm that improves the locality of a loop nest by transforming the code via interchange, reversal, skewing and tiling. The loop transformation algorithm is based on two concepts: a mathematical formulation of reuse and locality, and a loop transformation theory that unifi ..."
Abstract - Cited by 804 (16 self) - Add to MetaCart
that unifies the various transforms as unimodular matrix transformations. The algorithm has been implemented in the SUIF (Stanford University Intermediate Format) compiler, and is successful in optimizing codes such as matrix multiplication, successive over-relaxation (SOR), LU decomposition without pivoting

Robust principal component analysis?

by Emmanuel J Candès , Xiaodong Li , Yi Ma , John Wright - Journal of the ACM, , 2011
"... Abstract This paper is about a curious phenomenon. Suppose we have a data matrix, which is the superposition of a low-rank component and a sparse component. Can we recover each component individually? We prove that under some suitable assumptions, it is possible to recover both the low-rank and the ..."
Abstract - Cited by 569 (26 self) - Add to MetaCart
Abstract This paper is about a curious phenomenon. Suppose we have a data matrix, which is the superposition of a low-rank component and a sparse component. Can we recover each component individually? We prove that under some suitable assumptions, it is possible to recover both the low

Sequential data assimilation with a nonlinear quasi-geostrophic model using Monte Carlo methods to forecast error statistics

by Geir Evensen - J. Geophys. Res , 1994
"... . A new sequential data assimilation method is discussed. It is based on forecasting the error statistics using Monte Carlo methods, a better alternative than solving the traditional and computationally extremely demanding approximate error covariance equation used in the extended Kalman filter. The ..."
Abstract - Cited by 800 (23 self) - Add to MetaCart
. A new sequential data assimilation method is discussed. It is based on forecasting the error statistics using Monte Carlo methods, a better alternative than solving the traditional and computationally extremely demanding approximate error covariance equation used in the extended Kalman filter

Biclustering algorithms for biological data analysis: a survey.

by Sara C Madeira , Arlindo L Oliveira - IEEE/ACM Transactions of Computational Biology and Bioinformatics, , 2004
"... Abstract A large number of clustering approaches have been proposed for the analysis of gene expression data obtained from microarray experiments. However, the results of the application of standard clustering methods to genes are limited. These limited results are imposed by the existence of a num ..."
Abstract - Cited by 481 (15 self) - Add to MetaCart
Abstract A large number of clustering approaches have been proposed for the analysis of gene expression data obtained from microarray experiments. However, the results of the application of standard clustering methods to genes are limited. These limited results are imposed by the existence of a
Next 10 →
Results 1 - 10 of 22,725
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University