Results 1  10
of
2,032,431
Uncertainty principles and ideal atomic decomposition
 IEEE Transactions on Information Theory
, 2001
"... Suppose a discretetime signal S(t), 0 t<N, is a superposition of atoms taken from a combined time/frequency dictionary made of spike sequences 1ft = g and sinusoids expf2 iwt=N) = p N. Can one recover, from knowledge of S alone, the precise collection of atoms going to make up S? Because every d ..."
Abstract

Cited by 588 (19 self)
 Add to MetaCart
Suppose a discretetime signal S(t), 0 t<N, is a superposition of atoms taken from a combined time/frequency dictionary made of spike sequences 1ft = g and sinusoids expf2 iwt=N) = p N. Can one recover, from knowledge of S alone, the precise collection of atoms going to make up S? Because every discretetime signal can be represented as a superposition of spikes alone, or as a superposition of sinusoids alone, there is no unique way of writing S as a sum of spikes and sinusoids in general. We prove that if S is representable as a highly sparse superposition of atoms from this time/frequency dictionary, then there is only one such highly sparse representation of S, and it can be obtained by solving the convex optimization problem of minimizing the `1 norm of the coe cients among all decompositions. Here \highly sparse " means that Nt + Nw < p N=2 where Nt is the number of time atoms, Nw is the number of frequency atoms, and N is the length of the discretetime signal.
A computational approach to edge detection
 IEEE Transactions on Pattern Analysis and Machine Intelligence
, 1986
"... AbstractThis paper describes a computational approach to edge detection. The success of the approach depends on the definition of a comprehensive set of goals for the computation of edge points. These goals must be precise enough to delimit the desired behavior of the detector while making minimal ..."
Abstract

Cited by 4621 (0 self)
 Add to MetaCart
. We use the criteria in numerical optimization to derive detectors for several common image features, including step edges. On specializing the analysis to step edges, we find that there is a natural uncertainty principle between detection and localization performance, which are the two main goals
Robust Uncertainty Principles: Exact Signal Reconstruction From Highly Incomplete Frequency Information
, 2006
"... This paper considers the model problem of reconstructing an object from incomplete frequency samples. Consider a discretetime signal and a randomly chosen set of frequencies. Is it possible to reconstruct from the partial knowledge of its Fourier coefficients on the set? A typical result of this pa ..."
Abstract

Cited by 2599 (51 self)
 Add to MetaCart
This paper considers the model problem of reconstructing an object from incomplete frequency samples. Consider a discretetime signal and a randomly chosen set of frequencies. Is it possible to reconstruct from the partial knowledge of its Fourier coefficients on the set? A typical result of this paper is as follows. Suppose that is a superposition of spikes @ Aa @ A @ A obeying @�� � A I for some constant H. We do not know the locations of the spikes nor their amplitudes. Then with probability at least I @ A, can be reconstructed exactly as the solution to the I minimization problem I aH @ A s.t. ” @ Aa ” @ A for all
Advances in Prospect Theory: Cumulative Representation of Uncertainty
 JOURNAL OF RISK AND UNCERTAINTY, 5:297323 (1992)
, 1992
"... We develop a new version of prospect theory that employs cumulative rather than separable decision weights and extends the theory in several respects. This version, called cumulative prospect theory, applies to uncertain as well as to risky prospects with any number of outcomes, and it allows differ ..."
Abstract

Cited by 1603 (12 self)
 Add to MetaCart
different weighting functions for gains and for losses. Two principles, diminishing sensitivity and loss aversion, are invoked to explain the characteristic curvature of the value function and the weighting functions. A review of the experimental evidence and the results of a new experiment confirm a
Ontologies: Principles, methods and applications
 KNOWLEDGE ENGINEERING REVIEW
, 1996
"... This paper is intended to serve as a comprehensive introduction to the emerging field concerned with the design and use of ontologies. We observe that disparate backgrounds, languages, tools, and techniques are a major barrier to effective communication among people, organisations, and/or software s ..."
Abstract

Cited by 570 (3 self)
 Add to MetaCart
This paper is intended to serve as a comprehensive introduction to the emerging field concerned with the design and use of ontologies. We observe that disparate backgrounds, languages, tools, and techniques are a major barrier to effective communication among people, organisations, and/or software systems. We show how the development and implementation of an explicit account of a shared understanding (i.e. an `ontology') in a given subject area, can improve such communication, which in turn, can give rise to greater reuse and sharing, interoperability, and more reliable software. After motivating their need, we clarify just what ontologies are and what purposes they serve. We outline a methodology for developing and evaluating ontologies, first discussing informal techniques, concerning such issues as scoping, handling ambiguity, reaching agreement and producing de nitions. We then consider the bene ts of and describe, a more formal approach. We revisit the scoping phase, and discuss the role of formal languages and techniques in the specification, implementation and evaluation of ontologies. Finally, we review the state of the art and practice in this emerging field,
Principled design of the modern web architecture
 ACM Trans. Internet Techn
"... The World Wide Web has succeeded in large part because its software architecture has been designed to meet the needs of an Internetscale distributed hypermedia system. The modern Web architecture emphasizes scalability of component interactions, generality of interfaces, independent deployment of c ..."
Abstract

Cited by 507 (14 self)
 Add to MetaCart
of the Hypertext Transfer Protocol and Uniform Resource Identifiers. We describe the software engineering principles guiding REST and the interaction constraints chosen to retain those principles, contrasting them to the constraints of other architectural styles. We then compare the abstract model to the currently
Principles and Practice in Second Language Acquisition
, 1982
"... This is the original version of Principles and Practice, as published in 1982, with only minor changes. It is gratifying to point out that many of the predictions made in this book were confirmed by subsequent research, for example, the superiority of comprehensibleinput based methods and sheltered ..."
Abstract

Cited by 717 (4 self)
 Add to MetaCart
This is the original version of Principles and Practice, as published in 1982, with only minor changes. It is gratifying to point out that many of the predictions made in this book were confirmed by subsequent research, for example, the superiority of comprehensibleinput based methods
Blind Signal Separation: Statistical Principles
, 2003
"... Blind signal separation (BSS) and independent component analysis (ICA) are emerging techniques of array processing and data analysis, aiming at recovering unobserved signals or `sources' from observed mixtures (typically, the output of an array of sensors), exploiting only the assumption of mut ..."
Abstract

Cited by 522 (4 self)
 Add to MetaCart
, to show how they stem from basic principles and how they relate to each other.
Wallflower: Principles and Practice of Background Maintenance
, 1999
"... Background maintenance is a frequent element of video surveillance systems. We develop Wallflower, a threecomponent system for background maintenance: the pixellevel component performs Wiener filtering to make probabilistic predictions of the expected background; the regionlevel component fills i ..."
Abstract

Cited by 468 (1 self)
 Add to MetaCart
Background maintenance is a frequent element of video surveillance systems. We develop Wallflower, a threecomponent system for background maintenance: the pixellevel component performs Wiener filtering to make probabilistic predictions of the expected background; the regionlevel component fills in homogeneous regions of foreground objects; and the framelevel component detects sudden, global changes in the image and swaps in better approximations of the background. We compare our system with 8 other background subtraction algorithms. Wallflower is shown to outperform previous algorithms by handling a greater set of the difficult situations that can occur. Finally, we analyze the experimental results and propose
Results 1  10
of
2,032,431