Results 1  10
of
3,175
Treadmarks: Shared memory computing on networks of workstations
 Computer
, 1996
"... TreadMarks supports parallel computing on networks of workstations by providing the application with a shared memory abstraction. Shared memory facilitates the transition from sequential to parallel programs. After identifying possible sources of parallelism in the code, most of the data structures ..."
Abstract

Cited by 487 (37 self)
 Add to MetaCart
TreadMarks supports parallel computing on networks of workstations by providing the application with a shared memory abstraction. Shared memory facilitates the transition from sequential to parallel programs. After identifying possible sources of parallelism in the code, most of the data structures
Direct BulkSynchronous Parallel Algorithms
 JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING
, 1992
"... We describe a methodology for constructing parallel algorithms that are transportable among parallel computers having different numbers of processors, different bandwidths of interprocessor communication and different periodicity of global synchronisation. We do this for the bulksynchronous paralle ..."
Abstract

Cited by 174 (27 self)
 Add to MetaCart
We describe a methodology for constructing parallel algorithms that are transportable among parallel computers having different numbers of processors, different bandwidths of interprocessor communication and different periodicity of global synchronisation. We do this for the bulksynchronous
Parallel Mining of Association Rules
 IEEE Transactions on Knowledge and Data Engineering
, 1996
"... We consider the problem of mining association rules on a sharednothing multiprocessor. We present three algorithms that explore a spectrum of tradeoffs between computation, communication, memory usage, synchronization, and the use of problemspecific information. The best algorithm exhibits near p ..."
Abstract

Cited by 325 (3 self)
 Add to MetaCart
We consider the problem of mining association rules on a sharednothing multiprocessor. We present three algorithms that explore a spectrum of tradeoffs between computation, communication, memory usage, synchronization, and the use of problemspecific information. The best algorithm exhibits near
BulkSynchronous Parallel Computing on the CELL Processor
"... Abstract. In order to ease programming of heterogeneous architectures with explicitly managed memory hierarchies such as the CELL processor, we propose a solution adopting the BSP model as implemented in the parallel programming language NestStep. This allows the programmer to write programs with a ..."
Abstract
 Add to MetaCart
well as their execution time is mostly dominated by calculations, and only a fraction is spent in the various parts of the NestStep runtime system library. The library also has a relatively small memory footprint in the SPE's software managed local memory. Key words: Cell processor, NestStep, bulksynchronous
M.: NestStepModelica: Mathematical Modeling and BulkSynchronous Parallel Simulation
 PARA06 Workshop on stateoftheart in scientific and parallel computing
"... Abstract. Many parallel computing applications are used for simulation of complex engineering applications and/or for visualization. To handle their complexity, there is a need for raising the level of abstraction in specifying such applications using high level mathematical modeling techniques, su ..."
Abstract

Cited by 3 (3 self)
 Add to MetaCart
. In this context, we propose to combine the abstraction power of Modelica with support for shared memory bulksynchronous parallel programming including nested parallelism (NestStepModelica), which is both flexible (can be mapped to many different parallel architectures) and simple (offers a shared address space
Scientific Computing on Bulk Synchronous Parallel Architectures
"... We theoretically and experimentally analyse the efficiency with which a wide range of important scientific computations can be performed on bulk synchronous parallel architectures. ..."
Abstract

Cited by 75 (16 self)
 Add to MetaCart
We theoretically and experimentally analyse the efficiency with which a wide range of important scientific computations can be performed on bulk synchronous parallel architectures.
IStructures: Data Structures for Parallel Computing
, 1989
"... It is difficult to achieve elegance, efficiency and parallelism simultaneously in functional programs that manipulate large data structures. We demonstrate this through careful analysis of program examples using three common functional datastructuring approaches lists using Cons and arrays us ..."
Abstract

Cited by 266 (4 self)
 Add to MetaCart
using Update (both finegrained operators), and arrays using make array (a "bulk" operator). We then present Istructures as an alternative, and show elegant, efficient and parallel solutions for the program examples in Id, a language with Istructures. The parallelism in Id is made
The Design and Analysis of BulkSynchronous Parallel Algorithms
, 1998
"... The model of bulksynchronous parallel (BSP) computation is an emerging paradigm of generalpurpose parallel computing. This thesis presents a systematic approach to the design and analysis of BSP algorithms. We introduce an extension of the BSP model, called BSPRAM, which reconciles sharedmemory s ..."
Abstract

Cited by 19 (2 self)
 Add to MetaCart
The model of bulksynchronous parallel (BSP) computation is an emerging paradigm of generalpurpose parallel computing. This thesis presents a systematic approach to the design and analysis of BSP algorithms. We introduce an extension of the BSP model, called BSPRAM, which reconciles shared
The parallel evaluation of general arithmetic expressions
 Journal of the ACM
, 1974
"... ABSTRACT. It is shown that arithmetic expressions with n> 1 variables and constants; operations of addition, multiplication, and division; and any depth of parenthesis nesting can be evaluated in time 4 log2n + 10(n 1)/p using p> 1 processors which can independently perform arithmetic operati ..."
Abstract

Cited by 259 (1 self)
 Add to MetaCart
ABSTRACT. It is shown that arithmetic expressions with n> 1 variables and constants; operations of addition, multiplication, and division; and any depth of parenthesis nesting can be evaluated in time 4 log2n + 10(n 1)/p using p> 1 processors which can independently perform arithmetic
Bulk Synchronous Parallel Scheduling of Uniform Dags
 EuroPar'96. Parallel Processing, Lecture Notes in Computer Science 1124
, 1996
"... . This paper addresses the dag scheduling problem, proposing the bulk synchronous parallel (BSP) model as a framework for the derivation of general purpose parallel computer schedules of uniform dags, i.e., of dags that stand for tightlynested loops with computable distance vectors. A general techn ..."
Abstract

Cited by 4 (4 self)
 Add to MetaCart
. This paper addresses the dag scheduling problem, proposing the bulk synchronous parallel (BSP) model as a framework for the derivation of general purpose parallel computer schedules of uniform dags, i.e., of dags that stand for tightlynested loops with computable distance vectors. A general
Results 1  10
of
3,175