• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 2,259
Next 10 →

Dryad: Distributed Data-Parallel Programs from Sequential Building Blocks

by Michael Isard, Mihai Budiu, Yuan Yu, Andrew Birrell, Dennis Fetterly - In EuroSys , 2007
"... Dryad is a general-purpose distributed execution engine for coarse-grain data-parallel applications. A Dryad applica-tion combines computational “vertices ” with communica-tion “channels ” to form a dataflow graph. Dryad runs the application by executing the vertices of this graph on a set of availa ..."
Abstract - Cited by 762 (27 self) - Add to MetaCart
Dryad is a general-purpose distributed execution engine for coarse-grain data-parallel applications. A Dryad applica-tion combines computational “vertices ” with communica-tion “channels ” to form a dataflow graph. Dryad runs the application by executing the vertices of this graph on a set

Fast Parallel Algorithms for Short-Range Molecular Dynamics

by Steve Plimpton - JOURNAL OF COMPUTATIONAL PHYSICS , 1995
"... Three parallel algorithms for classical molecular dynamics are presented. The first assigns each processor a fixed subset of atoms; the second assigns each a fixed subset of inter-atomic forces to compute; the third assigns each a fixed spatial region. The algorithms are suitable for molecular dyn ..."
Abstract - Cited by 653 (7 self) - Add to MetaCart
dynamics models which can be difficult to parallelize efficiently -- those with short-range forces where the neighbors of each atom change rapidly. They can be implemented on any distributed--memory parallel machine which allows for message--passing of data between independently executing processors

Active Messages: a Mechanism for Integrated Communication and Computation

by Thorsten Von Eicken, David E. Culler, Seth Copen Goldstein, Klaus Erik Schauser , 1992
"... The design challenge for large-scale multiprocessors is (1) to minimize communication overhead, (2) allow communication to overlap computation, and (3) coordinate the two without sacrificing processor cost/performance. We show that existing message passing multiprocessors have unnecessarily high com ..."
Abstract - Cited by 1054 (75 self) - Add to MetaCart
tremendous flexibility. Implementations on nCUBE/2 and CM-5 are described and evaluated using a split-phase shared-memory extension to C, Split-C. We further show that active messages are sufficient to implement the dynamically scheduled languages for which message driven machines were designed

Partitioning of Unstructured Problems for Parallel Processing

by Horst D. Simon , 1991
"... Many large scale computational problems are based on unstructured computational domains. Primary examples are unstructured grid calculations based on finite volume methods in computational fluid dynamics, or structural analysis problems based on finite element approximations. Here we will address th ..."
Abstract - Cited by 344 (16 self) - Add to MetaCart
the question of how to dis-tribute such unstructured computational domains over a large number of processors in a MIMD machine with distributed memory. A graph theoretical framework for these problems will be established. Based on this framework three decomposition algorithms will be introduced. In particular

Scheduling Memory Constrained Jobs on Distributed Memory Parallel Computers

by Cathy McCann, John Zahorjan - IN PROCEEDINGS OF INTERNATIONAL JOINT CONFERENCE ON MEASUREMENT AND MODELING OF COMPUTER SYSTEMS, ACM SIGMETRICS 95 AND PERFORMANCE 95 , 1995
"... While the parallel use of many processors is a major attraction of scalable multiprocessors for large applications, another important feature of such machines is the large amount of physical memory they make available. Despite these resources, truly large applications may be limited not only by the ..."
Abstract - Cited by 37 (0 self) - Add to MetaCart
address three primary questions in this work: 1. How can a parallel machine be multiprogrammed with minimal overhead when jobs have minimum memory requirements? 2. To what extent does the inability of an application to repartition its workload during runtime affect the choice of processor allocation

Scheduling Memory Constrained Jobs on Distributed Memory Parallel Computers

by Cathy Mccann Tera, Cathy Mccann, John Zahorjan , 1995
"... We consider the problem of multiprocessor scheduling of jobs whose memory requirements place lower bounds on the fraction of the machine required in order to execute. We address three primary questions in this work: 1. How can a parallel machine be multiprogrammed with minimal overhead when jobs h ..."
Abstract - Add to MetaCart
We consider the problem of multiprocessor scheduling of jobs whose memory requirements place lower bounds on the fraction of the machine required in order to execute. We address three primary questions in this work: 1. How can a parallel machine be multiprogrammed with minimal overhead when jobs

Using Processor Affinity in Loop Scheduling on Shared-Memory Multiprocessors

by Evangelos P. Markatos, Thomas J. Leblanc - IEEE Transactions on Parallel and Distributed Systems , 1994
"... Loops are the single largest source of parallelism in many applications. One way to exploit this parallelism is to execute loop iterations in parallel on different processors. Previous approaches to loop scheduling attempt to achieve the minimum completion time by distributing the workload as evenly ..."
Abstract - Cited by 157 (2 self) - Add to MetaCart
Loops are the single largest source of parallelism in many applications. One way to exploit this parallelism is to execute loop iterations in parallel on different processors. Previous approaches to loop scheduling attempt to achieve the minimum completion time by distributing the workload

Thread scheduling for multiprogrammed multiprocessors

by Nimar S. Arora, Robert D. Blumofe, C. Greg Plaxton - In Proceedings of the Tenth Annual ACM Symposium on Parallel Algorithms and Architectures (SPAA), Puerto Vallarta , 1998
"... We present a user-level thread scheduler for shared-memory multiprocessors, and we analyze its performance under multiprogramming. We model multiprogramming with two scheduling levels: our scheduler runs at user-level and schedules threads onto a fixed collection of processes, while below, the opera ..."
Abstract - Cited by 208 (3 self) - Add to MetaCart
We present a user-level thread scheduler for shared-memory multiprocessors, and we analyze its performance under multiprogramming. We model multiprogramming with two scheduling levels: our scheduler runs at user-level and schedules threads onto a fixed collection of processes, while below

Processor Scheduling In A Distributed-Memory Computing Environment

by Stephen W. Turner , 1995
"... In recent years, the development of large-scale distributed-memory computers has given the user community unprecedented levels of computing power. In order to effectively use the available computing power, processor scheduling algorithms have been developed that allow many users to share distribut ..."
Abstract - Add to MetaCart
In recent years, the development of large-scale distributed-memory computers has given the user community unprecedented levels of computing power. In order to effectively use the available computing power, processor scheduling algorithms have been developed that allow many users to share

A Distributed Resource Management Architecture that Supports Advance Reservations and Co-Allocation

by Ian Foster, Carl Kesselman, Craig Lee, Bob Lindell, Klara Nahrstedt, Alain Roy
"... The realization of end-to-end quality of service (QoS) guarantees in emerging network-based applications requires mechanisms that support first dynamic discovery and then advance or immediate reservation of resources that will often be heterogeneous in type and implementation and independently contr ..."
Abstract - Cited by 245 (26 self) - Add to MetaCart
controlled and administered.We propose the Globus Architecture for Reservation and Allocation (GARA) to address these four issues.GARA treats both reservations and computational elements such as processes, network flows, and memory blocks as first class entities, allowing them to be created, monitored
Next 10 →
Results 1 - 10 of 2,259
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University