Results 1  10
of
46,125
Inapproximability of combinatorial public projects
, 2008
"... We study the Combinatorial Public Project Problem (CPPP) in which n agents are assigned a subset of m resources of size k so as to maximize the social welfare. Combinatorial public projects are an abstraction of many resourceassignment problems (Internetrelated network design, elections, etc.). ..."
Abstract

Cited by 7 (5 self)
 Add to MetaCart
We study the Combinatorial Public Project Problem (CPPP) in which n agents are assigned a subset of m resources of size k so as to maximize the social welfare. Combinatorial public projects are an abstraction of many resourceassignment problems (Internetrelated network design, elections, etc
Near Optimal Signal Recovery From Random Projections: Universal Encoding Strategies?
, 2004
"... Suppose we are given a vector f in RN. How many linear measurements do we need to make about f to be able to recover f to within precision ɛ in the Euclidean (ℓ2) metric? Or more exactly, suppose we are interested in a class F of such objects— discrete digital signals, images, etc; how many linear m ..."
Abstract

Cited by 1513 (20 self)
 Add to MetaCart
Suppose we are given a vector f in RN. How many linear measurements do we need to make about f to be able to recover f to within precision ɛ in the Euclidean (ℓ2) metric? Or more exactly, suppose we are interested in a class F of such objects— discrete digital signals, images, etc; how many linear measurements do we need to recover objects from this class to within accuracy ɛ? This paper shows that if the objects of interest are sparse or compressible in the sense that the reordered entries of a signal f ∈ F decay like a powerlaw (or if the coefficient sequence of f in a fixed basis decays like a powerlaw), then it is possible to reconstruct f to within very high accuracy from a small number of random measurements. typical result is as follows: we rearrange the entries of f (or its coefficients in a fixed basis) in decreasing order of magnitude f  (1) ≥ f  (2) ≥... ≥ f  (N), and define the weakℓp ball as the class F of those elements whose entries obey the power decay law f  (n) ≤ C · n −1/p. We take measurements 〈f, Xk〉, k = 1,..., K, where the Xk are Ndimensional Gaussian
Distributed Computing in Practice: The Condor Experience
 Concurrency and Computation: Practice and Experience
, 2005
"... Since 1984, the Condor project has enabled ordinary users to do extraordinary computing. Today, the project continues to explore the social and technical problems of cooperative computing on scales ranging from the desktop to the worldwide computational grid. In this chapter, we provide the history ..."
Abstract

Cited by 542 (7 self)
 Add to MetaCart
Since 1984, the Condor project has enabled ordinary users to do extraordinary computing. Today, the project continues to explore the social and technical problems of cooperative computing on scales ranging from the desktop to the worldwide computational grid. In this chapter, we provide
The Proposition Bank: An Annotated Corpus of Semantic Roles
 Computational Linguistics
, 2005
"... The Proposition Bank project takes a practical approach to semantic representation, adding a layer of predicateargument information, or semantic role labels, to the syntactic structures of the Penn Treebank. The resulting resource can be thought of as shallow, in that it does not represent corefere ..."
Abstract

Cited by 536 (21 self)
 Add to MetaCart
The Proposition Bank project takes a practical approach to semantic representation, adding a layer of predicateargument information, or semantic role labels, to the syntactic structures of the Penn Treebank. The resulting resource can be thought of as shallow, in that it does not represent
FFTW: An Adaptive Software Architecture For The FFT
, 1998
"... FFT literature has been mostly concerned with minimizing the number of floatingpoint operations performed by an algorithm. Unfortunately, on presentday microprocessors this measure is far less important than it used to be, and interactions with the processor pipeline and the memory hierarchy have ..."
Abstract

Cited by 605 (4 self)
 Add to MetaCart
, with over 40 implementations of the FFT on 7 machines. Our tests show that FFTW's selfoptimizing approach usually yields significantly better performance than all other publicly available software. FFTW also compares favorably with machinespecific, vendoroptimized libraries. 1. INTRODUCTION
Fuzzy extractors: How to generate strong keys from biometrics and other noisy data. Technical Report 2003/235, Cryptology ePrint archive, http://eprint.iacr.org, 2006. Previous version appeared at EUROCRYPT 2004
 34 [DRS07] [DS05] [EHMS00] [FJ01] Yevgeniy Dodis, Leonid Reyzin, and Adam
, 2004
"... We provide formal definitions and efficient secure techniques for • turning noisy information into keys usable for any cryptographic application, and, in particular, • reliably and securely authenticating biometric data. Our techniques apply not just to biometric information, but to any keying mater ..."
Abstract

Cited by 532 (38 self)
 Add to MetaCart
if the input changes, as long as it remains reasonably close to the original. Thus, R can be used as a key in a cryptographic application. A secure sketch produces public information about its input w that does not reveal w, and yet allows exact recovery of w given another value that is close to w. Thus
How Iris Recognition Works
, 2003
"... Algorithms developed by the author for recognizing persons by their iris patterns have now been tested in six field and laboratory trials, producing no false matches in several million comparison tests. The recognition principle is the failure of a test of statistical independence on iris phase st ..."
Abstract

Cited by 495 (4 self)
 Add to MetaCart
structure encoded by multiscale quadrature wavelets. The combinatorial complexity of this phase information across different persons spans about 244 degrees of freedom and generates a discrimination entropy of about 3.2 bits/mm over the iris, enabling realtime decisions about personal identity
Mining Frequent Patterns without Candidate Generation: A FrequentPattern Tree Approach
 DATA MINING AND KNOWLEDGE DISCOVERY
, 2004
"... Mining frequent patterns in transaction databases, timeseries databases, and many other kinds of databases has been studied popularly in data mining research. Most of the previous studies adopt an Apriorilike candidate set generationandtest approach. However, candidate set generation is still co ..."
Abstract

Cited by 1700 (64 self)
 Add to MetaCart
Mining frequent patterns in transaction databases, timeseries databases, and many other kinds of databases has been studied popularly in data mining research. Most of the previous studies adopt an Apriorilike candidate set generationandtest approach. However, candidate set generation is still costly, especially when there exist a large number of patterns and/or long patterns. In this study, we propose a novel
frequentpattern tree
(FPtree) structure, which is an extended prefixtree
structure for storing compressed, crucial information about frequent patterns, and develop an efficient FPtree
based mining method, FPgrowth, for mining the complete set of frequent patterns by pattern fragment growth.
Efficiency of mining is achieved with three techniques: (1) a large database is compressed into a condensed,
smaller data structure, FPtree which avoids costly, repeated database scans, (2) our FPtreebased mining adopts
a patternfragment growth method to avoid the costly generation of a large number of candidate sets, and (3) a
partitioningbased, divideandconquer method is used to decompose the mining task into a set of smaller tasks for
mining confined patterns in conditional databases, which dramatically reduces the search space. Our performance
study shows that the FPgrowth method is efficient and scalable for mining both long and short frequent patterns,
and is about an order of magnitude faster than the Apriori algorithm and also faster than some recently reported
new frequentpattern mining methods
An introduction to Kolmogorov Complexity and its Applications: Preface to the First Edition
, 1997
"... This document has been prepared using the L a T E X system. We thank Donald Knuth for T E X, Leslie Lamport for L a T E X, and Jan van der Steen at CWI for online help. Some figures were prepared by John Tromp using the xpic program. The London Mathematical Society kindly gave permission to reproduc ..."
Abstract

Cited by 2143 (120 self)
 Add to MetaCart
This document has been prepared using the L a T E X system. We thank Donald Knuth for T E X, Leslie Lamport for L a T E X, and Jan van der Steen at CWI for online help. Some figures were prepared by John Tromp using the xpic program. The London Mathematical Society kindly gave permission to reproduce a long extract by A.M. Turing. The Indian Statistical Institute, through the editor of Sankhy¯a, kindly gave permission to quote A.N. Kolmogorov. We gratefully acknowledge the financial support by NSF Grant DCR8606366, ONR Grant N0001485k0445, ARO Grant DAAL0386K0171, the Natural Sciences and Engineering Research Council of Canada through operating grants OGP0036747, OGP046506, and International Scientific Exchange Awards ISE0046203, ISE0125663, and NWO Grant NF 62376. The book was conceived in late Spring 1986 in the Valley of the Moon in Sonoma County, California. The actual writing lasted on and off from autumn 1987 until summer 1993. One of us [PV] gives very special thanks to his lovely wife Pauline for insisting from the outset on the significance of this enterprise. The Aiken Computation Laboratory of Harvard University, Cambridge, Massachusetts, USA; the Computer Science Department of York University, Ontario, Canada; the Computer Science Department of the University xii of Waterloo, Ontario, Canada; and CWI, Amsterdam, the Netherlands provided the working environments in which this book could be written. Preface to the Second Edition
Counterexampleguided Abstraction Refinement
, 2000
"... We present an automatic iterative abstractionrefinement methodology in which the initial abstract model is generated by an automatic analysis of the control structures in the program to be verified. Abstract models may admit erroneous (or "spurious") counterexamples. We devise new symb ..."
Abstract

Cited by 848 (71 self)
 Add to MetaCart
We present an automatic iterative abstractionrefinement methodology in which the initial abstract model is generated by an automatic analysis of the control structures in the program to be verified. Abstract models may admit erroneous (or "spurious") counterexamples. We devise new symbolic techniques which analyze such counterexamples and refine the abstract model correspondingly.
Results 1  10
of
46,125