Results 1  10
of
24,825
A hardcore predicate for all oneway functions
 In Proceedings of the Twenty First Annual ACM Symposium on Theory of Computing
, 1989
"... Abstract rity of f. In fact, for inputs (to f*) of practical size, the pieces effected by f are so small A central tool in constructing pseudorandom that f can be inverted (and the “hardcore” generators, secure encryption functions, and bit computed) by exhaustive search. in other areas are “hardc ..."
Abstract

Cited by 440 (5 self)
 Add to MetaCart
*, which has oneway function f(p, x) = (p,g(x)). The rea hardcore predicate. The construction ap sult extends to multiple (up to the logarithm plies the original f to many small pieces of of security) such bits and to any distribution the input to f * just to get one “hardcore ” on the z’s for which f
RSVP: A New Resource Reservation Protocol
, 1993
"... Whe origin of the RSVP protocol can be traced back to 1991, when a team of network researchers, including myself, started playing with a number of packet scheduling algorithms on the DARTNET (DARPA Testbed NETwork), a network testbed made of open source, workstationbased routers. Because scheduling ..."
Abstract

Cited by 1005 (25 self)
 Add to MetaCart
used by applications to communicate their requirements to the network in a robust and efficient way, independent of the specific requirements.” It has been more than 10 years since the original idea
Bilateral Filtering for Gray and Color Images
, 1998
"... tomasi @ cs.stanford.edu Bilateral filtering smooths images while preserving edges, by means of a nonlinear combination of nearby image values. The method is noniterative, local, and simple. It combines gray levels or colors based on both their geometric closeness and their photometric similariv, a ..."
Abstract

Cited by 1156 (2 self)
 Add to MetaCart
, and prefers near values to distant values in both domain and range. In contrast with filters that operate on the three bands of a color image separately, a bilateral filter can enforce the perceptual metric underlying the CIELab color space, and smooth colors and preserve edges in a way that is tuned
A review of image denoising algorithms, with a new one
 SIMUL
, 2005
"... The search for efficient image denoising methods is still a valid challenge at the crossing of functional analysis and statistics. In spite of the sophistication of the recently proposed methods, most algorithms have not yet attained a desirable level of applicability. All show an outstanding perf ..."
Abstract

Cited by 508 (6 self)
 Add to MetaCart
is proven to be asymptotically optimal under a generic statistical image model. The denoising performance of all considered methods are compared in four ways; mathematical: asymptotic order of magnitude of the method noise under regularity assumptions; perceptualmathematical: the algorithms artifacts
The particel swarm: Explosion, stability, and convergence in a multidimensional complex space
 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTION
"... The particle swarm is an algorithm for finding optimal regions of complex search spaces through interaction of individuals in a population of particles. Though the algorithm, which is based on a metaphor of social interaction, has been shown to perform well, researchers have not adequately explained ..."
Abstract

Cited by 852 (10 self)
 Add to MetaCart
’s convergence tendencies. Some results of the particle swarm optimizer, implementing modifications derived from the analysis, suggest methods for altering the original algorithm in ways that eliminate problems and increase the optimization power of the particle swarm
Scalespace and edge detection using anisotropic diffusion
 IEEE Transactions on Pattern Analysis and Machine Intelligence
, 1990
"... AbstractThe scalespace technique introduced by Witkin involves generating coarser resolution images by convolving the original image with a Gaussian kernel. This approach has a major drawback: it is difficult to obtain accurately the locations of the “semantically meaningful ” edges at coarse sca ..."
Abstract

Cited by 1887 (1 self)
 Add to MetaCart
AbstractThe scalespace technique introduced by Witkin involves generating coarser resolution images by convolving the original image with a Gaussian kernel. This approach has a major drawback: it is difficult to obtain accurately the locations of the “semantically meaningful ” edges at coarse
Collaborative plans for complex group action
, 1996
"... The original formulation of SharedPlans by B. Grosz and C. Sidner ( 1990) was developed to provide a model of collaborative planning in which it was not necessary for one agent to have intentionsto toward an act of a different agent. Unlike other contemporaneous approaches (J.R. Searle, 1990), this ..."
Abstract

Cited by 543 (30 self)
 Add to MetaCart
The original formulation of SharedPlans by B. Grosz and C. Sidner ( 1990) was developed to provide a model of collaborative planning in which it was not necessary for one agent to have intentionsto toward an act of a different agent. Unlike other contemporaneous approaches (J.R. Searle, 1990
Ktheory for operator algebras
 Mathematical Sciences Research Institute Publications
, 1998
"... p. XII line5: since p. 12: I blew this simple formula: should be α = −〈ξ, η〉/〈η, η〉. p. 2 I.1.1.4: The RieszFischer Theorem is often stated this way today, but neither Riesz nor Fischer (who worked independently) phrased it in terms of completeness of the orthogonal system {e int}. If [a, b] is a ..."
Abstract

Cited by 558 (0 self)
 Add to MetaCart
p. XII line5: since p. 12: I blew this simple formula: should be α = −〈ξ, η〉/〈η, η〉. p. 2 I.1.1.4: The RieszFischer Theorem is often stated this way today, but neither Riesz nor Fischer (who worked independently) phrased it in terms of completeness of the orthogonal system {e int}. If [a, b
Loopy belief propagation for approximate inference: An empirical study. In:
 Proceedings of Uncertainty in AI,
, 1999
"... Abstract Recently, researchers have demonstrated that "loopy belief propagation" the use of Pearl's polytree algorithm in a Bayesian network with loops can perform well in the context of errorcorrecting codes. The most dramatic instance of this is the near Shannonlimit performanc ..."
Abstract

Cited by 676 (15 self)
 Add to MetaCart
steady states it seems reasonable to try to find a way to com bine the two values. The simplest thing to do is to average them. Unfortunately, this gave very poor re sults, since the correct posteriors do not usually lie in the midpoint of the interval ( cf. 2More precisely, we found that with a
ReTiling Polygonal Surfaces
 Computer Graphics
, 1992
"... This paper presents an automatic method of creating surface models at several levels of detail from an original polygonal description of a given object. Representing models at various levels of detail is important for achieving high frame rates in interactive graphics applications and also for speed ..."
Abstract

Cited by 445 (3 self)
 Add to MetaCart
surface that is faithful to both the geometry and the topology of the original surface. Themain contributions of this paper are: 1) a robust method of connecting together new vertices over a surface, 2) a way of using an estimate of surface curvature to distribute more new vertices at regions of higher
Results 1  10
of
24,825