• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 11 - 20 of 2,259
Next 10 →

Multifrontal Parallel Distributed Symmetric and Unsymmetric Solvers

by P. R. Amestoy, I. S. Duff, J.-Y. L'Excellent , 1998
"... We consider the solution of both symmetric and unsymmetric systems of sparse linear equations. A new parallel distributed memory multifrontal approach is described. To handle numerical pivoting efficiently, a parallel asynchronous algorithm with dynamic scheduling of the computing tasks has been dev ..."
Abstract - Cited by 187 (30 self) - Add to MetaCart
We consider the solution of both symmetric and unsymmetric systems of sparse linear equations. A new parallel distributed memory multifrontal approach is described. To handle numerical pivoting efficiently, a parallel asynchronous algorithm with dynamic scheduling of the computing tasks has been

Evaluating MapReduce for multi-core and multiprocessor systems

by Colby Ranger, Ramanan Raghuraman, Arun Penmetsa, Gary Bradski, Christos Kozyrakis - In HPCA ’07: Proceedings of the 13th International Symposium on High-Performance Computer Architecture , 2007
"... This paper evaluates the suitability of the MapReduce model for multi-core and multi-processor systems. MapReduce was created by Google for application development on data-centers with thousands of servers. It allows programmers to write functional-style code that is automatically parallelized and s ..."
Abstract - Cited by 256 (3 self) - Add to MetaCart
and scheduled in a distributed system. We describe Phoenix, an implementation of MapReduce for shared-memory systems that includes a programming API and an efficient runtime system. The Phoenix runtime automatically manages thread creation, dynamic task scheduling, data partitioning, and fault tolerance across

Clustering and Intra-Processor Scheduling for Explicitly-Parallel Programs on Distributed-Memory Systems

by Vibha A. Dixit-radiya, Dhabaleswar K. Panda, Vibha A. Dixit-radiya, Dhabaleswar K. P, Dhabaleswar K. P - In International Parallel Processing Symposium , 1994
"... Programs for distributed-memory systems are explicitly-parallel and comprise of a set of sequential tasks or processes that communicate via message-passing. The sequence of computation in each task together with the intermediate send and receive communication steps exhibit temporal behavior of the p ..."
Abstract - Cited by 7 (2 self) - Add to MetaCart
Programs for distributed-memory systems are explicitly-parallel and comprise of a set of sequential tasks or processes that communicate via message-passing. The sequence of computation in each task together with the intermediate send and receive communication steps exhibit temporal behavior

Processor Allocation in Multiprogrammed Distributed-Memory Parallel Computer Systems

by Vijay Naik, Sanjeev K. Setia, Mark S. Squillante, Dr. Vijay, K. Naik , 1997
"... : In this paper, we examine three general classes of space-sharing scheduling policies under a workload representative of large-scale scientific computing. These policies differ in the way processors are partitioned among the jobs as well as in the way jobs are prioritized for execution on the parti ..."
Abstract - Cited by 19 (4 self) - Add to MetaCart
: In this paper, we examine three general classes of space-sharing scheduling policies under a workload representative of large-scale scientific computing. These policies differ in the way processors are partitioned among the jobs as well as in the way jobs are prioritized for execution

A Parallel Hashed Oct-Tree N-Body Algorithm

by Michael S. Warren, John K. Salmon , 1993
"... We report on an efficient adaptive N-body method which we have recently designed and implemented. The algorithm computes the forces on an arbitrary distribution of bodies in a time which scales as N log N with the particle number. The accuracy of the force calculations is analytically bounded, and c ..."
Abstract - Cited by 199 (14 self) - Add to MetaCart
We report on an efficient adaptive N-body method which we have recently designed and implemented. The algorithm computes the forces on an arbitrary distribution of bodies in a time which scales as N log N with the particle number. The accuracy of the force calculations is analytically bounded

ScaLAPACK: A Portable Linear Algebra Library for Distributed Memory Computers -- Design Issues and Performance (Technical Paper)

by J. Choi, J. Demmel, I. Dhillon, J. Dongarra, S. Ostrouchov, A. Petitet, K. Stanley, D. Walker, R. C. Whaley , 1996
"... This paper outlines the content and performance of ScaLAPACK, a collection of mathematical software for linear algebra computations on distributed memory computers. The importance of developing standards for computational and message passing interfaces is discussed. We present the different componen ..."
Abstract - Cited by 170 (56 self) - Add to MetaCart
components and building blocks of ScaLAPACK. This paper outlines the difficulties inherent in producing correct codes for networks of heterogeneous processors. We define a theoretical model of parallel computers dedicated to linear algebra applications: the Distributed Linear Algebra Machine (DLAM

Communication Optimizations for Irregular Scientific Computations on Distributed Memory Architectures

by Raja Das, Mustafa Uysal, Joel Saltz, Yuan-shin Hwang - Journal of Parallel and Distributed Computing , 1993
"... This paper describes a number of optimizations that can be used to support the efficient execution of irregular problems on distributed memory parallel machines. These primitives (1) coordinate interprocessor data movement, (2) manage the storage of, and access to, copies of off-processor data, (3) ..."
Abstract - Cited by 144 (17 self) - Add to MetaCart
This paper describes a number of optimizations that can be used to support the efficient execution of irregular problems on distributed memory parallel machines. These primitives (1) coordinate interprocessor data movement, (2) manage the storage of, and access to, copies of off-processor data, (3

The impact of operating system scheduling policies and synchronization methods of performance of parallel applications

by Anoop Gupta, Andrew Tucker, Shigeru Urushibara - In SIGMETRICS , 1991
"... Shared-memory multiprocessors are frequently used as compute servers with multiple parallel applications executing at the same time. In such environments, the efficiency of a parrdlel ap-plication can be significantly affected by the operating system scheduling policy. In this paper, we use detailed ..."
Abstract - Cited by 139 (2 self) - Add to MetaCart
Shared-memory multiprocessors are frequently used as compute servers with multiple parallel applications executing at the same time. In such environments, the efficiency of a parrdlel ap-plication can be significantly affected by the operating system scheduling policy. In this paper, we use

Scheduling and Page Migration for Multiprocessor Compute Servers

by Rohit Chandra, Scott Devine, Ben Verghese, Anoop Gupta, Mendel Rosenblum - INTERNATIONAL CONFERENCE ON ARCHITECTURAL SUPPORT FOR PROGRAMMING LANGUAGES AND OPERATING SYSTEMS , 1994
"... Several cache-coherent shared-memory multiprocessors have been developed that are scalable and offer a very tight coupling between the processing resources. They are therefore quite attractive for use as compute servers for multiprogramming and parallel application workloads. Process scheduling and ..."
Abstract - Cited by 103 (5 self) - Add to MetaCart
Several cache-coherent shared-memory multiprocessors have been developed that are scalable and offer a very tight coupling between the processing resources. They are therefore quite attractive for use as compute servers for multiprogramming and parallel application workloads. Process scheduling

Sequoia: Programming the Memory Hierarchy

by Kayvon Fatahalian, Timothy J. Knight, Mike Houston, Mattan Erez, Daniel Reiter Horn, Larkhoon Leem , Ji Young Park, Manman Ren, Alex Aiken, William J. Dally, Pat Hanrahan , 2006
"... We present Sequoia, a programming language designed to facilitate the development of memory hierarchy aware parallel programs that remain portable across modern machines featuring different memory hierarchy configurations. Sequoia abstractly exposes hierarchical memory in the programming model and p ..."
Abstract - Cited by 156 (9 self) - Add to MetaCart
and provides language mechanisms to describe communication vertically through the machine and to localize computation to particular memory locations within it. We have implemented a complete programming system, including a compiler and runtime systems for Cell processor-based blade systems and distributed
Next 10 →
Results 11 - 20 of 2,259
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University