Results 1  10
of
13
Performance Models for the Processor Farm Paradigm
 IEEE Transactions on Parallel and Distributed Systems
, 1997
"... In this paper, we describe the design, implementation, and modeling of a runtime kernel to support the processor farm paradigm on multicomputers. We present a general topologyindependent framework for obtaining performance models to predict the performance of the startup, steadystate, and wind ..."
Abstract

Cited by 18 (0 self)
 Add to MetaCart
(Show Context)
In this paper, we describe the design, implementation, and modeling of a runtime kernel to support the processor farm paradigm on multicomputers. We present a general topologyindependent framework for obtaining performance models to predict the performance of the startup, steadystate, and winddown phases of a processor farm. An algorithm is described, which for any interconnection network determines a treestructured subnetwork that optimizes farm performance. The analysis technique is applied to the important case of kary tree topologies. The models are compared with the measured performance on a variety of topologies using both constant and varied task sizes. Index TermsParallel programming paradigms, performance evaluation, processor farm, tree networks, message passing architecture, network flow, masterslave.  F  1I NTRODUCTION HE major problems in parallel computation revolve around questions of ease of...
Clumps: A Candidate Model Of Efficient, General Purpose Parallel Computation
, 1994
"... A new model of parallel computation is proposed, CLUMPS (Campbell's Lenient, Unified Model of Parallel Systems). This is composed of an abstract machine with an associated cost model, and aims to be more portable, reflective of costs, expressible and encouraging of more efficient implementation ..."
Abstract

Cited by 10 (6 self)
 Add to MetaCart
A new model of parallel computation is proposed, CLUMPS (Campbell's Lenient, Unified Model of Parallel Systems). This is composed of an abstract machine with an associated cost model, and aims to be more portable, reflective of costs, expressible and encouraging of more efficient implementations of algorithms than other existing models. It is shown that each basic parallel architecture class can congruently perform each other's computations, but the congruent simulation of each other's communication is not generally possible (where for a simulation to be congruent the simulation costs on the target architecture are asymptotically equivalent to the implementation costs on the native architectures). This is reflected in the CLUMPS abstract machine through its flexibility in terms of program control and memory access. The congruence requirement is relaxed so that though strict congruence may not be achieved according to the above definition, communication costs are reflectively accounted ...
Parallel cellular automata: a model program for computational science, Concurrency: Practice and Experience 5
, 1993
"... We develop a model program for parallel execution of cellular automata on a multicomputer. The model program is then adapted for simulation of forest fires and numerical solution of Laplace's equation for stationary heat fl.ow. The performance of the parallel program is analyzed and measured o ..."
Abstract

Cited by 8 (0 self)
 Add to MetaCart
We develop a model program for parallel execution of cellular automata on a multicomputer. The model program is then adapted for simulation of forest fires and numerical solution of Laplace's equation for stationary heat fl.ow. The performance of the parallel program is analyzed and measured on a Computing Surface configured as a matrix of transputers with distributed memory.
Further towards a unification of parallel architecture classes
 Department of Computer Science, University fo York
, 1996
"... ..."
Unbalanced Computations onto a Transputer Grid
, 1994
"... Many applications exists that are characterised by having some "core" function code repeatedly applied over all the elements of a compound data structure or over all the elements of an input data stream. Here, a technique is discussed that allows parallel implementations of these applicati ..."
Abstract

Cited by 2 (2 self)
 Add to MetaCart
Many applications exists that are characterised by having some "core" function code repeatedly applied over all the elements of a compound data structure or over all the elements of an input data stream. Here, a technique is discussed that allows parallel implementations of these applications to be derived that achieve high performance and efficiency in the machine resource usage. The technique is especially effective when the computational load of each application subtasks requires a variable amount of time to be computed. Some results concerning the usage of these techniques on a transputer based machine are discussed here along with the technical details of the implementation schemas that have been used. 1 Introduction Many scientific and nonscientific applications can be parallelised in an easy way. We refer to the computations that are characterised by a large amount of time spent in repeatedly computing the same functions/procedures/statements over different input data sets a...
Machine independent Analytical models for cost evaluation of templatebased programs
, 1996
"... Structured parallel programming is one of the possible solutions to exploit Programmability, Portability and Performance in the parallel programming world. The power of this approach stands in the possibility to build an optimizing template based compiler using low time complexity algorithms. I ..."
Abstract

Cited by 1 (0 self)
 Add to MetaCart
Structured parallel programming is one of the possible solutions to exploit Programmability, Portability and Performance in the parallel programming world. The power of this approach stands in the possibility to build an optimizing template based compiler using low time complexity algorithms. In order to optimize the code, this compiler needs formulas that describe the performance of language constructs over the target architecture. We propose a set of parameters able to describe current parallel systems and build deterministic analytical models for basic forms of parallelism. The analytical model describes construct performance in a parametric way.This can be done by knowing that the compiler exploits a templatebased support and giving template implementors guidelines to follow to make actual implementation perform as predicted. ACMCR Subject Classification: Keyword and phrases: Skeletons, performance modeling, parallel languages, template based compilers Machine...
Parallelizing A Model Checker
"... A new parallel algorithm that performes CTL model checking is given and its performance evaluated. It is based on a matrix representation of the computational model, which is distributed among the processing elements. The algorithm has been implemented on a Transputerbased architecture Meiko CS/1. ..."
Abstract
 Add to MetaCart
(Show Context)
A new parallel algorithm that performes CTL model checking is given and its performance evaluated. It is based on a matrix representation of the computational model, which is distributed among the processing elements. The algorithm has been implemented on a Transputerbased architecture Meiko CS/1. Experimental results and analytical modeling are presented here that show the "good" parallel behaviour of the algorithm. Keywords: Model checking, MIMD, data parallelism, Transputer, temporal logic. 1 Introduction In the last years it has been shown that temporal logics could be a useful tool for concurrent system specification and verification. One of this logics, namely CTL, seems to be particularly suitable for the purpose of automatic verification. In fact in [1,4] efficient algorithms for the socalled model checking problem have been given and its practical application demonstrated. The model checking problem for CTL can be stated as follows: given a CTL formula and a finite comp...
Adaptable Planner Primitives for RealWorld Robotic Applications
"... With increased processor speed and improved robotic and AI technology, researchers are beginning to design programs that can behave intelligently and interact in the real world. A large increase in processing power has come from parallel machines, but taking advantage of this power is challenging. I ..."
Abstract
 Add to MetaCart
With increased processor speed and improved robotic and AI technology, researchers are beginning to design programs that can behave intelligently and interact in the real world. A large increase in processing power has come from parallel machines, but taking advantage of this power is challenging. In this paper we address the issues in designing planners for realtime AI and robotic applications, and provide guiding principles. These principles were designed to minimize the difference between the new realtime model and the standard offline model. Applying these principles yields a betterstructured application, easier design and implementation, and improved performance. The focus of the paper is on a design methodology for implementing effective planners in realworld applications. Using Ephor (our runtime environment), and applying the described planner principles, we demonstrate improved performance in a realworld shepherding application. 1
Parallel Cellular Automata: A Model
"... We develop a generic program for parallel execution of cellular automata on a multicomputer. The generic program is then adapted for simulation of a forest fire and numerical solution of Laplace’s equation for stationary heat flow. The performance of the parallel program is analyzed and measured on ..."
Abstract
 Add to MetaCart
(Show Context)
We develop a generic program for parallel execution of cellular automata on a multicomputer. The generic program is then adapted for simulation of a forest fire and numerical solution of Laplace’s equation for stationary heat flow. The performance of the parallel program is analyzed and measured on a Computing Surface configured as a matrix of transputers with distributed memory. 1
Model Programs for Computational Science: A Programming Methodology
"... We describe a programming methodology for computational science based on programming paradigms for multicomputers. Each paradigm is a class of algorithms that have the same control structure. For every paradigm, a general parallel program is developed. The general program is then used to derive two ..."
Abstract
 Add to MetaCart
(Show Context)
We describe a programming methodology for computational science based on programming paradigms for multicomputers. Each paradigm is a class of algorithms that have the same control structure. For every paradigm, a general parallel program is developed. The general program is then used to derive two or more model programs, which solve specific problems in science and engineering. These programs have been tested on a Computing Surface and published with every detail open to scrutiny. We explain the steps involved in developing model programs and conclude that the study of programming paradigms provides an architectural vision of parallel scientific computing. 1