Results 1  10
of
28
Sharedmemory mutual exclusion: Major research trends since
 Distributed Computing
, 1986
"... * Exclusion: At most one process executes its critical section at any time. ..."
Abstract

Cited by 51 (6 self)
 Add to MetaCart
(Show Context)
* Exclusion: At most one process executes its critical section at any time.
Adaptive and Efficient Algorithms for Lattice Agreement and Renaming
 SIAM J. Comput
, 1998
"... In a sharedmemory system, n independent asynchronous processes, with distinct names in the range {0, ..., N  1}, communicate by reading and writing to shared registers. An algorithm is waitfree if a process completes its execution regardless of the behavior of other processes. This paper consider ..."
Abstract

Cited by 23 (8 self)
 Add to MetaCart
In a sharedmemory system, n independent asynchronous processes, with distinct names in the range {0, ..., N  1}, communicate by reading and writing to shared registers. An algorithm is waitfree if a process completes its execution regardless of the behavior of other processes. This paper considers waitfree algorithms whose complexity adjusts to the level of contention in the system: An algorithm is adaptive (to total contention) if its step complexity depends only on the actual number of active processes, k; this number is unknown in advance and may change in different executions of the algorithm. Adaptive algorithms are presented for two important decision problems, lattice agreement and (6k  1)renaming; the step complexity of both algorithms is O(k log k). An interesting component of the (6k  1)renaming algorithm is an O(N) algorithm for (2k  1)renaming; this improves on the best previously known (2k  1)renaming algorithm, which has O(Nnk) s...
Algorithms adaptive to point contention
 Journal of the ACM
, 2003
"... Abstract. This article introduces the sieve, a novel building block that allows to adapt to the number of simultaneously active processes (the point contention) during the execution of an operation. We present an implementation of the sieve in which each sieve operation requires O(k log k) steps, wh ..."
Abstract

Cited by 19 (8 self)
 Add to MetaCart
Abstract. This article introduces the sieve, a novel building block that allows to adapt to the number of simultaneously active processes (the point contention) during the execution of an operation. We present an implementation of the sieve in which each sieve operation requires O(k log k) steps, where k is the point contention during the operation. The sieve is the cornerstone of the first waitfree algorithms that adapt to point contention using only read and write operations. Specifically, we present efficient algorithms for longlived renaming, timestamping and collecting information.
Efficient Adaptive Collect using Randomization
 PROC. OF THE INTL. SYMP. ON DISTRIBUTED COMPUTING (DISC
, 2004
"... An adaptive algorithm, whose step complexity adjusts to the number of active processes, is attractive for distributed systems with a highlyvariable number of processes. The cornerstone of many adaptive algorithms is an adaptive mechanism to collect uptodate information from all participating p ..."
Abstract

Cited by 17 (2 self)
 Add to MetaCart
(Show Context)
An adaptive algorithm, whose step complexity adjusts to the number of active processes, is attractive for distributed systems with a highlyvariable number of processes. The cornerstone of many adaptive algorithms is an adaptive mechanism to collect uptodate information from all participating processes. To date, all known collect algorithms either have nonlinear step complexity or they are impractical because of unrealistic memory overhead. This paper
A Simple Algorithmic Characterization of Uniform Solvability (Extended Abstract)
 Proceedings of the 43rd Annual IEEE Symposium on Foundations of Computer Science (FOCS 2002
, 2002
"... The HerlihyShavit (HS) conditions characterizing the solvability of asynchronous tasks over n processors have been a milestone in the development of the theory of distributed computing. Yet, they were of no help when researcher sought algorithms that do not depend on n. To help in this pursuit we i ..."
Abstract

Cited by 11 (6 self)
 Add to MetaCart
(Show Context)
The HerlihyShavit (HS) conditions characterizing the solvability of asynchronous tasks over n processors have been a milestone in the development of the theory of distributed computing. Yet, they were of no help when researcher sought algorithms that do not depend on n. To help in this pursuit we investigate the uniform solvability of an infinite uniform sequence of tasks T 0 , T 1 , T 2 , ..., where T i is a task over processors p 0 , p 1 , ..., p i , and T i extends T i1 . We say that such a sequence is uniformly solvable if there exit protocols to solve each T i and the protocol for T i extends the protocol for T i1 . This paper establishes that although each T i may be solvable, the uniform sequence is not necessarily uniformly solvable. We show this by proposing a novel uniform sequence of solvable tasks and proving that the sequence is not amenable to a uniform solution. We then extend the HS conditions for a task over n processors, to uniform solvability in a natural way. The technique we use to accomplish this is to generalize the alternative algorithmic proof, by Borowsky and Gafni, of the HS conditions, by showing that the infinite uniform sequence of task of Immediate Snapshots is uniformly solvable. A side benefit of the technique is a widely applicable methodology for the development of uniform protocols.
Lamport on Mutual Exclusion: 27 Years of Planting Seeds
 In 20th ACM Symposium on Principles of Distributed Computing
, 2001
"... Mutual exclusion is a topic that Leslie Lamport has returned to many times throughout his career. This article, which is being written in celebration of Lamport's sixtieth birthday, is an attempt to survey some of his many contributions to research on this topic. ..."
Abstract

Cited by 9 (0 self)
 Add to MetaCart
(Show Context)
Mutual exclusion is a topic that Leslie Lamport has returned to many times throughout his career. This article, which is being written in celebration of Lamport's sixtieth birthday, is an attempt to survey some of his many contributions to research on this topic.
Uniform solvability with a finite number of mwmr registers
 In Proceedings of the 17th International Conference on Distributed Computing
, 2003
"... Abstract. This paper introduces a new interesting research question concerning tasks. The weaktestandset task has a uniform solution that requires only two MultiWriter MultiReader (MWMR) registers. Recently it was shown that if we take the longlived version and require a step complexity that i ..."
Abstract

Cited by 7 (5 self)
 Add to MetaCart
(Show Context)
Abstract. This paper introduces a new interesting research question concerning tasks. The weaktestandset task has a uniform solution that requires only two MultiWriter MultiReader (MWMR) registers. Recently it was shown that if we take the longlived version and require a step complexity that is adaptive to interval contention then, like mutual exclusion, no solution with finitely many MWMR registers is possible. Here we show that there are simple tasks which provably cannot be solved uniformly with finitely many MWMR registers. This opens up the research question of when a task is uniformly solvable using only finitely many MWMR registers. 1
An O(1) RMRs leader election algorithm
 In Proc. ACM PODC 2006
, 2006
"... The leader election problem is a fundamental coordination problem. We present leader election algorithms for multiprocessor systems where processes communicate by reading and writing shared memory asynchronously, and do not fail. In particular, we consider the cachecoherent (CC) and distributed shar ..."
Abstract

Cited by 6 (3 self)
 Add to MetaCart
(Show Context)
The leader election problem is a fundamental coordination problem. We present leader election algorithms for multiprocessor systems where processes communicate by reading and writing shared memory asynchronously, and do not fail. In particular, we consider the cachecoherent (CC) and distributed shared memory (DSM) models of such systems. We present leader election algorithms that perform a constant number of remote memory references (RMRs) in the worst case. Our algorithms use splitterlike objects [6, 9] in a novel way, by organizing active processes into teams that share work. As there is an Ω(log n) lower bound on the RMR complexity of mutual exclusion for n processes using reads and writes only [10], our result separates the mutual exclusion and leader election problems in terms of RMR complexity in both the CC and DSM models. Our result also implies that any algorithm using reads, writes and onetime testandset objects can be simulated by an algorithm using reads and writes with only a constant blowup of the RMR complexity; proving this is easy in the CC model, but presents subtle challenges in
Computing in the presence of timing failures
 In Proceedings of theInternational Conference on Distributed Computing Systems (DCS
, 2007
"... Timing failures refer to a situation where the environment in which a system operates does not behave as expected regarding the timing assumptions, that is, the timing constraints are not met. In the immense body of work on the designing faulttolerant systems, the type of failures that are usually ..."
Abstract

Cited by 5 (1 self)
 Add to MetaCart
(Show Context)
Timing failures refer to a situation where the environment in which a system operates does not behave as expected regarding the timing assumptions, that is, the timing constraints are not met. In the immense body of work on the designing faulttolerant systems, the type of failures that are usually considered are, process failures, link failures, messages loss and memory failures; and it is usually (implicitly) assumed that there are no timing failures. In this paper we investigate the ability to recover automatically from transient timing failures. We introduce and formally define the concept of algorithms that are resilient to timing failures, and demonstrate the importance of the new concept by presenting consensus and mutual exclusion algorithms, using atomic registers only, that are resilient to timing failures.