Results 1  10
of
57,078
Hamming Approximation of NP Witnesses
, 2013
"... Given a satisfiable 3SAT formula, how hard is it to find an assignment to the variables that has Hamming distance at most n/2 to a satisfying assignment? More generally, consider any polynomialtime verifier for any NPcomplete language. A d(n)Hammingapproximation algorithm for the verifier is o ..."
Abstract

Cited by 4 (0 self)
 Add to MetaCart
is one that, given any member x of the language, outputs in polynomial time a string a with Hamming distance at most d(n) to some witness w, where (x,w) is accepted by the verifier. Previous results have shown that, if P=NP, every NPcomplete language has a verifier for which there is no (n/2 − n 2/3+δ)Hammingapproximation
A Threshold of ln n for Approximating Set Cover
 JOURNAL OF THE ACM
, 1998
"... Given a collection F of subsets of S = f1; : : : ; ng, set cover is the problem of selecting as few as possible subsets from F such that their union covers S, and max kcover is the problem of selecting k subsets from F such that their union has maximum cardinality. Both these problems are NPhar ..."
Abstract

Cited by 778 (5 self)
 Add to MetaCart
hard. We prove that (1 \Gamma o(1)) ln n is a threshold below which set cover cannot be approximated efficiently, unless NP has slightly superpolynomial time algorithms. This closes the gap (up to low order terms) between the ratio of approximation achievable by the greedy algorithm (which is (1 \Gamma
A Guided Tour to Approximate String Matching
 ACM COMPUTING SURVEYS
, 1999
"... We survey the current techniques to cope with the problem of string matching allowing errors. This is becoming a more and more relevant issue for many fast growing areas such as information retrieval and computational biology. We focus on online searching and mostly on edit distance, explaining t ..."
Abstract

Cited by 584 (38 self)
 Add to MetaCart
We survey the current techniques to cope with the problem of string matching allowing errors. This is becoming a more and more relevant issue for many fast growing areas such as information retrieval and computational biology. We focus on online searching and mostly on edit distance, explaining the problem and its relevance, its statistical behavior, its history and current developments, and the central ideas of the algorithms and their complexities. We present a number of experiments to compare the performance of the different algorithms and show which are the best choices according to each case. We conclude with some future work directions and open problems.
The English noun phrase in its sentential aspect
, 1987
"... This dissertation is a defense of the hypothesis that the noun phrase is headed by afunctional element (i.e., \nonlexical " category) D, identi ed with the determiner. In this way, the structure of the noun phrase parallels that of the sentence, which is headed by In (ection), under assump ..."
Abstract

Cited by 509 (4 self)
 Add to MetaCart
phrases. The problem of capturing this dual aspect of the Possing construction is heightened by current restrictive views of Xbar theory, which, in particular, rule out the obvious structure for Possing, [NP NP VPing], by virtue of its exocentricity. Consideration of languages in which nouns, even
Graphical models, exponential families, and variational inference
, 2008
"... The formalism of probabilistic graphical models provides a unifying framework for capturing complex dependencies among random variables, and building largescale multivariate statistical models. Graphical models have become a focus of research in many statistical, computational and mathematical fiel ..."
Abstract

Cited by 800 (26 self)
 Add to MetaCart
all be understood in terms of exact or approximate forms of these variational representations. The variational approach provides a complementary alternative to Markov chain Monte Carlo as a general source of approximation methods for inference in largescale statistical models.
Parameterized Complexity
, 1998
"... the rapidly developing systematic connections between FPT and useful heuristic algorithms  a new and exciting bridge between the theory of computing and computing in practice. The organizers of the seminar strongly believe that knowledge of parameterized complexity techniques and results belongs ..."
Abstract

Cited by 1218 (75 self)
 Add to MetaCart
the rapidly developing systematic connections between FPT and useful heuristic algorithms  a new and exciting bridge between the theory of computing and computing in practice. The organizers of the seminar strongly believe that knowledge of parameterized complexity techniques and results belongs into the toolkit of every algorithm designer. The purpose of the seminar was to bring together leading experts from all over the world, and from the diverse areas of computer science that have been attracted to this new framework. The seminar was intended as the rst larger international meeting with a specic focus on parameterized complexity, and it hopefully serves as a driving force in the development of the eld. 1 We had 49 participants from Australia, Canada, India, Israel, New Zealand, USA, and various European countries. During the workshop 25 lectures were given. Moreover, one night session was devoted to open problems and Thursday was basically used for problem discussion
The lexical nature of syntactic ambiguity resolution
 Psychological Review
, 1994
"... Ambiguity resolution is a central problem in language comprehension. Lexical and syntactic ambiguities are standardly assumed to involve different types of knowledge representations and be resolved by different mechanisms. An alternative account is provided in which both types of ambiguity derive fr ..."
Abstract

Cited by 556 (23 self)
 Add to MetaCart
Ambiguity resolution is a central problem in language comprehension. Lexical and syntactic ambiguities are standardly assumed to involve different types of knowledge representations and be resolved by different mechanisms. An alternative account is provided in which both types of ambiguity derive from aspects of lexical representation and are resolved by the same processing mechanisms. Reinterpreting syntactic ambiguity resolution as a form of lexical ambiguity resolution obviates the need for special parsing principles to account for syntactic interpretation preferences, reconciles a number of apparently conflicting results concerning the roles of lexical and contextual information in sentence processing, explains differences among ambiguities in terms of ease of resolution, and provides a more unified account of language comprehension than was previously available. One of the principal goals for a theory of language compre third section we consider processing issues: how information is hension is to explain how the reader or listener copes with a processed within the mental lexicon and how contextual inforpervasive ambiguity problem. Languages are structured at mation can influence processing. The central processing mechmultiple levels simultaneously, including lexical, phonological, anism we invoke is the constraint satisfaction process that has morphological, syntactic, and text or discourse levels. At any been realized in interactiveactivation models (e.g., Elman &
Random Oracles are Practical: A Paradigm for Designing Efficient Protocols
, 1995
"... We argue that the random oracle model  where all parties have access to a public random oracle  provides a bridge between cryptographic theory and cryptographic practice. In the paradigm we suggest, a practical protocol P is produced by first devising and proving correct a protocol P R for the ..."
Abstract

Cited by 1643 (75 self)
 Add to MetaCart
We argue that the random oracle model  where all parties have access to a public random oracle  provides a bridge between cryptographic theory and cryptographic practice. In the paradigm we suggest, a practical protocol P is produced by first devising and proving correct a protocol P R for the random oracle model, and then replacing oracle accesses by the computation of an "appropriately chosen" function h. This paradigm yields protocols much more efficient than standard ones while retaining many of the advantages of provable security. We illustrate these gains for problems including encryption, signatures, and zeroknowledge proofs.
Mean shift, mode seeking, and clustering
 IEEE Transactions on Pattern Analysis and Machine Intelligence
, 1995
"... AbstractMean shift, a simple iterative procedure that shifts each data point to the average of data points in its neighborhood, is generalized and analyzed in this paper. This generalization makes some kmeans like clustering algorithms its special cases. It is shown that mean shift is a modeseeki ..."
Abstract

Cited by 620 (0 self)
 Add to MetaCart
AbstractMean shift, a simple iterative procedure that shifts each data point to the average of data points in its neighborhood, is generalized and analyzed in this paper. This generalization makes some kmeans like clustering algorithms its special cases. It is shown that mean shift is a modeseeking process on a surface constructed with a “shadow ” kernel. For Gaussian kernels, mean shift is a gradient mapping. Convergence is studied for mean shift iterations. Cluster analysis is treated as a deterministic problem of finding a fixed point of mean shift that characterizes the data. Applications in clustering and Hough transform are demonstrated. Mean shift is also considered as an evolutionary strategy that performs multistart global optimization. Index TermsMean shift, gradient descent, global optimization, Hough transform, cluster analysis, kmeans clustering. I.
Results 1  10
of
57,078