• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 11 - 20 of 1,571
Next 10 →

OpenMP on Networks of Workstations

by Honghui Lu Y, Y. Charlie Hu Z, Willy Zwaenepoel Z
"... We describe an implementation of a sizable subset of OpenMP on networks of workstations (NOWs). By extending the availability ofOpenMPtoNOWs, we overcome one of its primary drawbacks compared to MPI, namely lack of portability toenvironments other than hardware shared memory machines. In order to su ..."
Abstract - Add to MetaCart
of an OpenMP subset on a NOW, and compare them with hand-coded software DSM and MPI results for the same applications on the same platform. We use ve applications (ASCI Sweep3d, NAS 3D-FFT, SPLASH-2 Water, QSORT, and TSP) exhibiting various styles of parallelization, including pipelined execution, data

On Implementing MPI-IO Portably and with High Performance

by Rajeev Thakur, William Gropp, Ewing Lusk - In Proceedings of the 6th Workshop on I/O in Parallel and Distributed Systems , 1999
"... We discuss the issues involved in implementing MPI-IO portably on multiple machines and file systems and also achieving high performance. One way to implement MPI-IO portably is to implement it on top of the basic Unix I/O functions (open, lseek, read, write, and close), which are themselves portabl ..."
Abstract - Cited by 196 (20 self) - Add to MetaCart
We discuss the issues involved in implementing MPI-IO portably on multiple machines and file systems and also achieving high performance. One way to implement MPI-IO portably is to implement it on top of the basic Unix I/O functions (open, lseek, read, write, and close), which are themselves

The OpenMP Implementation of NAS Parallel Benchmarks and its Performance

by H. Jin, H. Jin, M. Frumkin, M. Frumkin, J. Yan, J. Yan , 1999
"... As the new ccNUMA architecture became popular in recent years, parallel programming with compiler directives on these machines has evolved to accommodate new needs. In this study, we examine the effectiveness of OpenMP directives for parallelizing the NAS Parallel Benchmarks. Implementation details ..."
Abstract - Cited by 167 (4 self) - Add to MetaCart
will be discussed and performance will be compared with the MPI implementation. We have demonstrated that OpenMP can achieve very good results for parallelization on a shared memory system, but effective use of memory and cache is very important.

PETSc users manual

by Satish Balay, Kris Buschelman, Victor Eijkhout, Dinesh Kaushik, Matt Knepley, Lois Curfman Mcinnes, Barry Smith, Hong Zhang - ANL-95/11 - Revision 2.1.0, Argonne National Laboratory , 2001
"... tract W-31-109-Eng-38. 2 This manual describes the use of PETSc for the numerical solution of partial differential equations and related problems on high-performance computers. The Portable, Extensible Toolkit for Scientific Compu-tation (PETSc) is a suite of data structures and routines that provid ..."
Abstract - Cited by 282 (20 self) - Add to MetaCart
that provide the building blocks for the implemen-tation of large-scale application codes on parallel (and serial) computers. PETSc uses the MPI standard for all message-passing communication. PETSc includes an expanding suite of parallel linear, nonlinear equation solvers and time integrators that may be used

A Comparison of MPI and OpenMP Implementations of a Finite Element Analysis Code

by Michael Bane, Rainer Keller, Michael Pettipher, Ian Smith , 2000
"... In this paper we describe the steps involved, the effort required and the performance achieved in both MPI and OpenMP implementations of a Fortran 90 finite element analysis code on an SGI Origin2000 using the MIPSpro compiler. We demonstrate that a working OpenMP version is easier to write, and the ..."
Abstract - Cited by 2 (0 self) - Add to MetaCart
In this paper we describe the steps involved, the effort required and the performance achieved in both MPI and OpenMP implementations of a Fortran 90 finite element analysis code on an SGI Origin2000 using the MIPSpro compiler. We demonstrate that a working OpenMP version is easier to write

Hybrid (OpenMP and MPI) Parallelization of MFIX: A Multiphase CFD Code for Modeling Fluidized Beds

by Sreekanth Pannala
"... We describe the effort and experience in generating a hybrid parallel version of MFIX (Multiphase Flow with Interphase eXchanges), a code for modeling reactive multiphase flow in fluidized beds. The code uses portable OpenMP and MPI in a unified source code. The resulting parallel code has been port ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
We describe the effort and experience in generating a hybrid parallel version of MFIX (Multiphase Flow with Interphase eXchanges), a code for modeling reactive multiphase flow in fluidized beds. The code uses portable OpenMP and MPI in a unified source code. The resulting parallel code has been

Techniques supporting threadprivate in OpenMP

by Xavier Martorell, Marc Gonzàlez, Alex Duran, Jairo Balart, Roger Ferrer, Eduard Ayguadé, Jesús Labarta - In Proceedings of the 11th International Workshop on High-Level Parallel Programming Models and Supportive , 2006
"... This paper presents the alternatives available to support threadprivate data in OpenMP and evaluates them. We show how current compilation systems rely on custom techniques for implementing thread-local data. But in fact the ELF binary specification currently supports data sections that become threa ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
only to ease porting MPI codes to OpenMP. 1. Introduction and

Performance analysis of large-scale OpenMP and hybrid MPI/OpenMP applications with VampirNG

by Vampir Ng, Holger Brunst, Bernd Mohr - in Proceedings of the First International Workshop on OpenMP (IWOMP 2005 , 2005
"... Abstract. This paper presents a tool setup for comprehensive eventbased performance analysis of large-scale openmp and hybrid openmp/ mpi applications. The kojak framework is used for portable code instrumentation and automatic analysis while the new Vampir NG infrastructure serves as generic visual ..."
Abstract - Cited by 11 (0 self) - Add to MetaCart
Abstract. This paper presents a tool setup for comprehensive eventbased performance analysis of large-scale openmp and hybrid openmp/ mpi applications. The kojak framework is used for portable code instrumentation and automatic analysis while the new Vampir NG infrastructure serves as generic

MPI versus MPI+OpenMP on the IBM SP for the NAS Benchmarks

by Franck Cappello, Daniel Etiemble , 2000
"... The hybrid memory model of clusters of multiprocessors raises two issues: programming model and performance. Many parallel programs have been written by using the MPI standard. To evaluate the pertinence of hybrid models for existing MPI codes, we compare a unified model (MPI) and a hybrid one (Open ..."
Abstract - Cited by 48 (0 self) - Add to MetaCart
The hybrid memory model of clusters of multiprocessors raises two issues: programming model and performance. Many parallel programs have been written by using the MPI standard. To evaluate the pertinence of hybrid models for existing MPI codes, we compare a unified model (MPI) and a hybrid one

Dual-Level Parallelism Exploitation with OpenMP

by In Coastal Ocean, Marc González, Eduard Ayguadé, Xavier Martorell, Jesús Labarta, Phu V. Luong - in Coastal Ocean Circulation Modeling”, International Workshop on OpenMP: Experiences and Implementations (WOMPEI 2002 , 2002
"... Two alternative dual-level parallel implementations of the Multiblock Grid Princeton Ocean Model (MGPOM) are compared in this paper. The first one combines the use of two programming paradigms: message passing with the Message Passing Interface (MPI) and shared memory with OpenMP (version called MPI ..."
Abstract - Add to MetaCart
MPI-OpenMP); the second uses only OpenMP (version called OpenMP-Only). MGPOM is a multiblock grid code that enables the exploitation of two levels of parallelism.
Next 10 →
Results 11 - 20 of 1,571
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University