Results 1  10
of
70
Proof verification and hardness of approximation problems
 IN PROC. 33RD ANN. IEEE SYMP. ON FOUND. OF COMP. SCI
, 1992
"... We show that every language in NP has a probablistic verifier that checks membership proofs for it using logarithmic number of random bits and by examining a constant number of bits in the proof. If a string is in the language, then there exists a proof such that the verifier accepts with probabilit ..."
Abstract

Cited by 704 (45 self)
 Add to MetaCart
We show that every language in NP has a probablistic verifier that checks membership proofs for it using logarithmic number of random bits and by examining a constant number of bits in the proof. If a string is in the language, then there exists a proof such that the verifier accepts with probability 1 (i.e., for every choice of its random string). For strings not in the language, the verifier rejects every provided “proof " with probability at least 1/2. Our result builds upon and improves a recent result of Arora and Safra [6] whose verifiers examine a nonconstant number of bits in the proof (though this number is a very slowly growing function of the input length). As a consequence we prove that no MAX SNPhard problem has a polynomial time approximation scheme, unless NP=P. The class MAX SNP was defined by Papadimitriou and Yannakakis [82] and hard problems for this class include vertex cover, maximum satisfiability, maximum cut, metric TSP, Steiner trees and shortest superstring. We also improve upon the clique hardness results of Feige, Goldwasser, Lovász, Safra and Szegedy [42], and Arora and Safra [6] and shows that there exists a positive ɛ such that approximating the maximum clique size in an Nvertex graph to within a factor of N ɛ is NPhard.
A formal analysis and taxonomy of task allocation in multirobot systems
 Int’l. J. of Robotics Research
"... Despite more than a decade of experimental work in multirobot systems, important theoretical aspects of multirobot coordination mechanisms have, to date, been largely untreated. To address this issue, we focus on the problem of multirobot task allocation (MRTA). Most work on MRTA has been ad hoc ..."
Abstract

Cited by 182 (4 self)
 Add to MetaCart
Despite more than a decade of experimental work in multirobot systems, important theoretical aspects of multirobot coordination mechanisms have, to date, been largely untreated. To address this issue, we focus on the problem of multirobot task allocation (MRTA). Most work on MRTA has been ad hoc and empirical, with many coordination architectures having been proposed and validated in a proofofconcept fashion, but infrequently analyzed. With the goal of bringing objective grounding to this important area of research, we present a formal study of MRTA problems. A domainindependent taxonomy of MRTA problems is given, and it is shown how many such problems can be viewed as instances of other, wellstudied, optimization problems. We demonstrate how relevant theory from operations research and combinatorial optimization can be used for analysis and greater understanding of existing approaches to task allocation, and to show how the same theory can be used in the synthesis of new approaches. KEY WORDS—task allocation, multirobot systems, coordination, utility 1.
The tracker: a threat to statistical database security
 ACM Trans. Database Syst
, 1979
"... The query programs of certain databases report raw statistics for query sets, which are groups of records specified implicitly by a characteristic formula. The raw statistics include query set size and sums of powers of values in the query set. Many users and designers believe that the individual re ..."
Abstract

Cited by 50 (2 self)
 Add to MetaCart
The query programs of certain databases report raw statistics for query sets, which are groups of records specified implicitly by a characteristic formula. The raw statistics include query set size and sums of powers of values in the query set. Many users and designers believe that the individual records will remain confidential as long as query programs refuse to report the statistics of query sets which are too small. It is shown that the compromise of small query sets can in fact almost always be accomplished with the help of characteristic formulas called trackers. Schlorer’s individual tracker is reviewed, it is derived from known characteristics of a given individual and permits deducing additional characteristics he may have. The general tracker is introduced: It permits calculating statistics for arbitrary query sets, without requiring preknowledge of anything in the database. General trackers always exist if there are enough distinguishable classes of individuals in the database, in which case the trackers have a simple form. Almost all databases have a general tracker, and general trackers are almost always easy to find. Security is not guaranteed by the lack of a general tracker.
Polynomial time approximation algorithms for machine scheduling: Ten open problems
 Journal of Scheduling
, 1999
"... We discuss what we consider to be the ten most vexing open questions in the area of polynomial time approximation algorithms for NPhard deterministic machine scheduling
problems. We summarize what is known on these problems, we discuss related results, and we provide pointers to the literature.
..."
Abstract

Cited by 36 (2 self)
 Add to MetaCart
We discuss what we consider to be the ten most vexing open questions in the area of polynomial time approximation algorithms for NPhard deterministic machine scheduling
problems. We summarize what is known on these problems, we discuss related results, and we provide pointers to the literature.
A Tabu Search Approach to Task Scheduling on Heterogeneous Processors under Precedence Constraints
, 1994
"... Parallel programs may be represented as a set of interrelated sequential tasks. When multiprocessors are used to execute such programs, the parallel portion of the application can be speeded up by an appropriate allocation of processors to the tasks of the application. Given a parallel application d ..."
Abstract

Cited by 33 (9 self)
 Add to MetaCart
Parallel programs may be represented as a set of interrelated sequential tasks. When multiprocessors are used to execute such programs, the parallel portion of the application can be speeded up by an appropriate allocation of processors to the tasks of the application. Given a parallel application defined by a task precedence graph, the goal of task scheduling (or processor assignment) is thus the minimization of the makespan of the application. In a heterogeneous multiprocessor system, task scheduling consists in determining which tasks will be assigned to each processor, as well as the execution order of the tasks assigned to each processor. In this work, we apply the tabu search metaheuristic to the solution of the task scheduling problem on a heterogeneous multiprocessor environment under precedence constraints. The topology of the Mean Value Analysis solution package for product form queueing networks is used as the framework for performance evaluation. We show that tabu search ob...
To weight or not to weight: where is the question?
 Proceedings of the 4th IEEE Israel Symposium on Theory of Computing and Systems
, 1996
"... We investigate the approximability properties of several weighted problems, by comparing them with the respective unweighted problems. For an appropriate (and very general) definition of niceness, we show that if a nice weighted problem is hard to approximate within r, then its polynomially bounded ..."
Abstract

Cited by 27 (7 self)
 Add to MetaCart
We investigate the approximability properties of several weighted problems, by comparing them with the respective unweighted problems. For an appropriate (and very general) definition of niceness, we show that if a nice weighted problem is hard to approximate within r, then its polynomially bounded weighted version is hard to approximate within r \Gamma o(1). Then we turn our attention to specific problems, and we show that the unweighted
Polynomial Time Approximation Schemes for ClassConstrained Packing Problems
 Proc. of Workshop on Approximation Algorithms
, 1999
"... . We consider variants of the classic bin packing and multiple knapsack problems, in which sets of items of different classes (colors) need to be placed in bins; the items may have different sizes and values. Each bin has a limited capacity, and a bound on the number of distinct classes of items ..."
Abstract

Cited by 26 (6 self)
 Add to MetaCart
. We consider variants of the classic bin packing and multiple knapsack problems, in which sets of items of different classes (colors) need to be placed in bins; the items may have different sizes and values. Each bin has a limited capacity, and a bound on the number of distinct classes of items it can hold. In the classconstrained multiple knapsack (CCMK) problem, our goal is to maximize the total value of packed items, whereas in the classconstrained binpacking (CCBP), we seek to minimize the number of (identical) bins, needed for packing all the items. We give a polynomial time approximation scheme (PTAS) for CCMK and a dual PTAS for CCBP. We also show that the 01 classconstrained knapsack admits a fully polynomial time approximation scheme, even when the number of distinct colors of items depends on the input size. Finally, we introduce the generalized classconstrained packing problem (GCCP), where each item may have more than one color. We show that GCCP is APX...
NearOptimal Dynamic Task Scheduling of Independent CoarseGrained Tasks onto a Computational Grid
 Proc. ICPP 2003
, 2003
"... problems is makespan. However, on a computational grid, the 2nd optimal makespan may be much longer than the optimal makespan because the computing power of a grid varies over time. So, if the performance measure is makespan, there is no approximation algorithm in general for scheduling onto a grid. ..."
Abstract

Cited by 21 (1 self)
 Add to MetaCart
problems is makespan. However, on a computational grid, the 2nd optimal makespan may be much longer than the optimal makespan because the computing power of a grid varies over time. So, if the performance measure is makespan, there is no approximation algorithm in general for scheduling onto a grid. In this paper, a novel criterion of a schedule is proposed. The proposed criterion is called total processor cycle consumption, which is the total number of instructions the grid could compute until the completion time of the schedule. Moreover, for the criterion, this paper gives a (1 + )approximation algorithm for scheduling n independent coarsegrained tasks with the same length onto a grid with m processors. The proposed algorithm does not use any prediction information on the performance of underlying resources. This result implies a nontrivial result that the computing power consumed by a parameter sweep application can be limited in such a case within (1 + ) times that required by an optimal schedule, regardless how the speed of each processor varies over time.
On weighted vs unweighted versions of combinatorial optimization problems
 Information and Computation
, 2001
"... We investigate the approximability properties of several weighted problems, by comparing them with the respective unweighted problems. For an appropriate (and very general) definition of niceness, we show that if a nice weighted problem is hard to approximate within r, then its polynomially bounded ..."
Abstract

Cited by 18 (2 self)
 Add to MetaCart
We investigate the approximability properties of several weighted problems, by comparing them with the respective unweighted problems. For an appropriate (and very general) definition of niceness, we show that if a nice weighted problem is hard to approximate within r, then its polynomially bounded weighted version is hard to approximate within r − o(1). Then we turn our attention to specific problems, and we show that the unweighted versions of Min Vertex Cover, Min Sat, Max Cut, Max DiCut, Max 2Sat, and Max Exact kSat are exactly as hard to approximate as their weighted versions. We note in passing that Min Vertex Cover is exactly as hard to approximate as Min Sat. In order to prove the reductions for Max 2Sat, Max Cut, Max DiCut, and Max E3Sat we introduce the new notion of “mixing ” set and we give an explicit construction of such sets. These reductions give new nonapproximability results for these problems. 1