• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Comparing the openmp, mpi, and hybrid programming paradigms on an smp cluster. In (2003)

by G Jost, H Jin, D an Mey, F F Hatay
Add To MetaCart

Tools

Sorted by:
Results 11 - 13 of 13

Vice Provost for Academic and International Programs ACKNOWLEDGMENTS

by Weirong Zhu, Weirong Zhu, Guang R. Gao, Ph. D, Gonzalo R. Arce, Ph. D, Eric W. Kaler, Ph. D, Firstly I
"... would like to thank Dr. Guang R. Gao for his advisement and support during the last two years. His insight and methodology on the research problems always teach me a lot. I really appreciate his comments and valuable guidance provided in the development of my thesis. I would also like to thank Dr. K ..."
Abstract - Add to MetaCart
would like to thank Dr. Guang R. Gao for his advisement and support during the last two years. His insight and methodology on the research problems always teach me a lot. I really appreciate his comments and valuable guidance provided in the development of my thesis. I would also like to thank Dr. Kevin Theobold, who guided me to the area of parallel system and computer architecture and taught me lots of computing skills during past three years. Also, I like to thank Yanwei Niu, Dr. Jizhu Lu, Chuan Shen, who worked with me for developing parallel system and parallel applications. One of important outcomes of our work { parallel HMMPFAM will be shown in this thesis. In addition, I would like to acknowledge Dr. Ziang Hu, Dr. Clement Leung, Yuan Zhang for helping to proof read and comment on my thesis at various stages of development. This research is funded in part by NSF, under the NGS grant 0103723, DOE, grant number DE-FC02-01ER25503. I also thank other US Federal agencies and
(Show Context)

Citation Context

...59][60] compares the performance of NAS benchmarks implemented by MPI versus hybrid MPI/OpenMP on dierent platforms. A comparison of OpenMP, MPI and hybrid programming on SMP cluster is presented in =-=[61]-=-. And Rabenseifner [62] discussed the performance problems and chances of hybrid programming. Hybrid programming can achieve good performance on SMP clusters, however, large amount 47 of programming e...

Recommended Citation

by Jonathan Lee Bentz, Jonathan Lee Bentz , 2006
"... Hybrid programming in high performance scientific computing ..."
Abstract - Add to MetaCart
Hybrid programming in high performance scientific computing
(Show Context)

Citation Context

... hybrid code is to optimize both thesinter and intra node communication of data. A number of studies have been done combiningsMPI and OpenMP to implement a so-called hybrid or multi-level parallelism =-=[3, 4, 5, 6, 7, 8]-=-.sThe most common hybrid model is the so-called "master/worker" model. One MPI process issexecuted per node (irrespective of the number of physical processors per node) and that processsspawns a maste...

GPU Computing for Parallel Local Search Metaheuristic Algorithms

by Thé Van Luong, Nouredine Melab, El-ghazali Talbi - IEEE TRANSACTIONS ON COMPUTERS , 2012
"... Local search metaheuristics (LSMs) are efficient methods for solving complex problems in science and industry. They allow significantly to reduce the size of the search space to be explored and the search time. Nevertheless, the resolution time remains prohibitive when dealing with large problem ins ..."
Abstract - Add to MetaCart
Local search metaheuristics (LSMs) are efficient methods for solving complex problems in science and industry. They allow significantly to reduce the size of the search space to be explored and the search time. Nevertheless, the resolution time remains prohibitive when dealing with large problem instances. Therefore, the use of GPU-based massively parallel computing is a major complementary way to speed up the search. However, GPU computing for LSMs is rarely investigated in the literature. In this paper, we introduce a new guideline for the design and implementation of effective LSMs on GPU. Very efficient approaches are proposed for CPU-GPU data transfer optimization, thread control, mapping of neighboring solutions to GPU threads and memory management. These approaches have been experimented using four well-known combinatorial and continuous optimization problems and four GPU configurations. Compared to a CPU-based execution, accelerations up to ×80 are reported for the large combinatorial problems and up to ×240 for a continuous problem. Finally, extensive experiments demonstrate the strong potential of GPU-based LSMs compared to cluster or grid-based parallel architectures.
(Show Context)

Citation Context

...penMP/MPI version has been produced to take advantage of both multi-core and distributed environments. Such a combination has widely proved in the past its effectiveness for multi-level architectures =-=[25]-=-. The PPP using a neighborhood based on a Hamming distance of two is considered on the two architectures. A Myri10G gigabit ethernet connects the different machines of the COWs. For the workstations d...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University