• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 1,571
Next 10 →

1. OpenMP and MPI

by Steven Gottlieb, Sonali Tamhankar , 2000
"... A trend in high performance computers that is becoming increasingly popular is the use of symmetric multiprocessing (SMP) rather than the older paradigm of MPP. MPI codes that ran and scaled well on MPP machines can often be run on an SMP machine using the vendor’s version of MPI. However, this appr ..."
Abstract - Add to MetaCart
to be able to use OpenMP parallelism on the node, and MPI between nodes. We describe the challenges of converting MILC MPI code to using a second level of OpenMP parallelism, and benchmarks on IBM and Sun computers.

Development of Mixed Mode MPI / OpenMP Applications

by Lorna Smith, Mark Bull - In WOMPAT 2000 , 2000
"... MPI / OpenMP mixed mode codes could potentially offer the most effective parallelisation strategy for an SMP cluster, as well as allowing the different characteristics of both paradigms to be exploited to give the best performance on a single SMP. This paper discusses the implementation, development ..."
Abstract - Cited by 37 (0 self) - Add to MetaCart
MPI / OpenMP mixed mode codes could potentially offer the most effective parallelisation strategy for an SMP cluster, as well as allowing the different characteristics of both paradigms to be exploited to give the best performance on a single SMP. This paper discusses the implementation

Intra node parallelization of MPI programs with OpenMP

by Franck Cappello, Olivier Richard , 1998
"... The availability of multiprocessors and high performance networks offer the opportunity to construct CLUMPs (Cluster of Multiprocessors) and use them as paxallel computing platforms. The main distinctive feature of the CLUMP axchitecture over the usual paxallel computers is its hybrid memory model ..."
Abstract - Cited by 3 (0 self) - Add to MetaCart
movements inside the CLUMP 3) to limit the effort of the programmer while ensuring the portability of the codes on a wide vaxiety of CLUMP configurations. We investigate an approach based on the MPI and OpenMP standaxds. The approach consists in the intra-node paxallelization of the MPI programs

Detecting threadsafety violations in hybrid openmp/mpi programs.

by Hongyi Ma , Liqiang Wang , Krishanthan Krishnamoorthy - In Proceedings of the 2015 IEEE International Conference on Cluster Computing, , 2015
"... Abstract-We propose an approach by integrating static and dynamic program analyses to detect threadsafety violations in hybrid MPI/OpenMP programs. We innovatively transform the thread-safety violation problems to race conditions problems. In our approach, the static analysis identifies a list of M ..."
Abstract - Cited by 1 (1 self) - Add to MetaCart
of MPI calls related to thread-safety violations, then replaces them with our own MPI wrappers, which involve accesses to some specific shared variables. The static analysis avoids instrumenting unrelated code, which significantly reduces runtime overhead. In the dynamic analysis, both happen

Parallelizing Particle-In-Cell Codes with OpenMP and MPI

by Nils Magnus Larsgård, Supervisor Anne, Cathrine Elster , 2007
"... This thesis searches for the best configuration of OpenMP/MPI for optimal performance. We will run a parallel simulation-application on a modern supercomputer to measure the effects of different configurations of the mixed code. After analyzing the performance of different configurations from a hard ..."
Abstract - Add to MetaCart
hardware point of view, we will propose a general model for estimating overhead in OpenMP loops. We will parallelize a physics simulation to do large scale simulations as efficient as possible. We will look at typical physics simulation codes to parallelize and optimize them with OpenMP and MPI in a mixed

Prototyping the automatic generation of MPI code from OpenMP programs in GCC

by Daniel Millot, Christian Parrot
"... Abstract. Multiprocessor architectures comprising various memory organizations and communi-cation schemes are now widely deployed. Hence, powerful programming tools are expressly needed in order to fully exploit the computation capabilities of such architectures. Classically, a parallel program must ..."
Abstract - Add to MetaCart
implementation in GCC, providing code trans-formations in the GIMPLE representation and compare them to those necessary for transforming a shared-memory program (annotated with OpenMP pragmas) to a message-passing MPI-based program. In order to test some ideas, we have developed a limited prototype. The result

OpenMP optimisation of a parallel MPI CFD code

by Fabrice Mathey, Philippe Blaise, Philippe Kloos
"... Introduction Despite the progress in numerical algorithm and computer power, reliable CFD prediction of complex flows as turbulent flows is still a challenge. Nu- merical turbulent prediction (Large Eddy Simulation, see [7] for more details) computes the three-dimensional, unsteady turbulent flow d ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
Introduction Despite the progress in numerical algorithm and computer power, reliable CFD prediction of complex flows as turbulent flows is still a challenge. Nu- merical turbulent prediction (Large Eddy Simulation, see [7] for more details) computes the three-dimensional, unsteady turbulent flow directly by use of the Navier-Stokes equations, with a model utilized only for the small scales of motion that are unresolved by the computational mesh (subgrid-scale model). The application of LES to industrial flows is highly attractive as the present increase in computing power of computers brings the necessary computational res-sources within the reach of industrial research and design groups. However, many situations still require High-Performance Computers. This requirement is due to the use of fine meshes with a huge number of points. A typical simulation the requires around 106 to 107 grid points. In turn, this implies a small size of the time step, increasing the number of iterations

Open MPI: Goals, concept, and design of a next generation MPI implementation

by Edgar Gabriel, Graham E. Fagg, George Bosilca, Thara Angskun, Jack J. Dongarra, Jeffrey M. Squyres, Vishal Sahay, Prabhanjan Kambadur, Brian Barrett, Andrew Lumsdaine, Ralph H. Castain, David J. Daniel, Richard L. Graham, Timothy S. Woodall - In Proceedings, 11th European PVM/MPI Users’ Group Meeting , 2004
"... Abstract. A large number of MPI implementations are currently available, each of which emphasize different aspects of high-performance computing or are intended to solve a specific research problem. The result is a myriad of incompatible MPI implementations, all of which require separate installatio ..."
Abstract - Cited by 223 (53 self) - Add to MetaCart
installation, and the combination of which present significant logistical challenges for end users. Building upon prior research, and influenced by experience gained from the code bases of the LAM/MPI, LA-MPI, and FT-MPI projects, Open MPI is an all-new, productionquality MPI-2 implementation

OpenMP on Networks of Workstations

by Honghui Lu , Y. Charlie Hu, Willy Zwaenepoel , 1998
"... We describe an implementation of a sizable subset of OpenMP on networks of workstations (NOWs). By extending the availability of OpenMP to NOWs, we overcome one of its primary drawbacks compared to MPI, namely lack of portability to environments other than hardware shared memory machines. In orde ..."
Abstract - Cited by 43 (6 self) - Add to MetaCart
We describe an implementation of a sizable subset of OpenMP on networks of workstations (NOWs). By extending the availability of OpenMP to NOWs, we overcome one of its primary drawbacks compared to MPI, namely lack of portability to environments other than hardware shared memory machines

Is OpenMP for Grids ?

by Rudolf Eigenmann, Jay Hoeflinger, Robert H. Kuhn, David Padua, Ayon Basumallik, Seung-Jai Min, Jiajing Zhu - in Proceedings of the International Parallel and Distributed Processing Symposium ( IPDPS ’02), Fort Lauderdale , 2002
"... This paper presents an overview of an ongoing NSFsponsored project for the study of runtime systems and compilers to support the development of efficient OpenMP parallel programs for distributed memory systems. The first part of the paper discusses a prototype compiler, now under development, that w ..."
Abstract - Cited by 3 (0 self) - Add to MetaCart
, that will accept OpenMP and will target TreadMarks, a Software Distributed Shared Memory System (SDSM), and Message-Passing Interface (MPI) library routines. A second part of the paper presents ideas for OpenMP extensions that enable the programmer to override the compiler whenever automatic methods fail
Next 10 →
Results 1 - 10 of 1,571
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University