Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/582034.582061acmconferencesArticle/Chapter ViewAbstractPublication PagesscConference Proceedingsconference-collections
Article

ORT: a communication library for orthogonal processor groups

Published: 10 November 2001 Publication History

Abstract

Many implementations on message-passing machines can benefit from an exploitation of mixed task and data parallelism. A suitable parallel programming model is a group-SPMD model, which requires a structuring of the processors into subsets and a partition of the program into multi-processor tasks. In this paper, we introduce a library support for the specification of message-passing programs in a group-SPMD style allowing different partitions in a single program. We describe the functionality and the implementation of the library functions and illustrate the library programming style with example programs. The examples show that the runtime on distributed memory machines can be considerably reduced by using the library.

References

[1]
S.B. Baden and S.J. Fink. A Programming Methodology for Dual-Tier Multicomputers. IEEE Transactions on Software Engineering, 26(3):212-226, 2000.
[2]
H. Bal and M. Haines. Approaches for Integrating Task and Data Parallelism. IEEE Concurrency, 6(3):74-84, July-August 1998.
[3]
C.J. Beckmann and C. Polychronopoulos. Microarchitecture Support for Dynamic Scheduling of Acyclic Task Graphs. Technical Report CSRD Report 1207, University of Illinois, 1992.
[4]
S.J. Fink. A Programming Model for Block-Structured Scientific Calculations on SMP Clusters. PhD thesis, University of California, San Diego, 1998.
[5]
I. Foster and K.M. Chandy. Fortran M: A Language for Modular Parallel Programming. Journal of Parallel and Distributed Computing, 25(1):24-35, April 1995.
[6]
I. Foster, M. Xu, B. Avalani, and A. Choudhary. A Compilation System That Integrates High Performance Fortran and Fortran M. In Proceedings 1994 Scalable High Performance Computing Conference, pages 293-300. IEEE Computer Society Press, 1994.
[7]
E. Hairer, S.P. Nørsett, and G. Wanner. Solving Ordinary Differential Equations I: Nonstiff Problems. Springer-Verlag, Berlin, 1993.
[8]
M. Hill, W. McColl, and D. Skillicorn. Questions and Answers about BSP. Scientific Programming, 6(3):249-274, 1997.
[9]
P. Joisha and P. Banerjee. PARADIGM (version 2.0): A New HPF Compilation System. In Proc. 1999 International Parallel Processing Symposium (IPPS'99), 1999.
[10]
S.R. Kohn and S.B. Baden. Irregular Coarse-Grain Data Parallelism under LPARX. Scientific Programming, 5:185-201, 1995.
[11]
W.F. McColl. Universal Computing. In Proceedings of the EuroPar'96, Springer LNCS 1123, pages 25-36, 1996.
[12]
J. Merlin, S.Baden, St. Fink, and B. Chapman. Multiple data parallelsim with HPF and KeLP. J. Future Generation Computer Science, 15(3):393-405, 1999.
[13]
S. Ramaswamy. Simultaneous Exploitation of Task and Data Parallelism in Regular Scientific Applications. PhD thesis, University of Illinois at Urbana-Champaign, 1996.
[14]
S. Ramaswamy, S. Sapatnekar, and P. Banerjee. A Framework for Exploiting Task and Data Parallelism on Distributed-Memory Multicomputers. IEEE Transactions on Parallel and Distributed Systems, 8(11):1098-1116, 1997.
[15]
T. Rauber and G. Rünger. Parallel Iterated Runge-Kutta Methods and Applications. International Journal of Super-computer Applications, 10(1):62-90, 1996.
[16]
T. Rauber and G. Rünger. Parallel Execution of Embedded and Iterated Runge-Kutta Methods. Concurrency: Practice and Experience, 11(7):367-385, 1999.
[17]
T. Rauber and G. Rünger. A Transformation Approach to Derive Efficient Parallel Implementations. IEEE Transactions on Software Engineering, 26(4):315-339, 2000.
[18]
D. Skillicorn and D. Talia. Models and languages for parallel computation. ACM Computing Surveys, 30(2):123-169, 1998.
[19]
J. Subhlok. Automatic Mapping of task and data Parallel programs for Efficient Execution on Multiprocessors. Technical Report CMU-CS-93-212, Carnegie Mellon University, 1993.
[20]
J. Subhlok and B. Yang. A New Model for Integrating Nested Task and Data Parallel Programming. In 8th ACM SIGPLAN Symp. on Principles & Practice of Parallel Programming, pages 1-12, 1997.
[21]
P.J. van der Houwen and B.P. Sommeijer. Parallel Iteration of high-order Runge-Kutta Methods with stepsize control. Journal of Computational and Applied Mathematics, 29:111-127, 1990.
[22]
G. Zhang, B. Carpenter, G.Fox, X. Li, and Y. Wen. A high level SPMD programming model: HPspmd and its Java language binding. Technical report, NPAC at Syracuse University, 1998.

Cited By

View all
  • (2018)Optimizing MPI collective communication by orthogonal structuresCluster Computing10.1007/s10586-006-9740-99:3(257-279)Online publication date: 24-Dec-2018
  • (2004)Improving the execution time of global communication operationsProceedings of the 1st conference on Computing frontiers10.1145/977091.977130(276-287)Online publication date: 14-Apr-2004
  • (2004)A source code analyzer for performance prediction18th International Parallel and Distributed Processing Symposium, 2004. Proceedings.10.1109/IPDPS.2004.1303333(262-269)Online publication date: 2004
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
SC '01: Proceedings of the 2001 ACM/IEEE conference on Supercomputing
November 2001
756 pages
ISBN:158113293X
DOI:10.1145/582034
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 10 November 2001

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. communication library
  2. group-SPMD
  3. mixed task and data parallelism

Qualifiers

  • Article

Conference

SC '01
Sponsor:

Acceptance Rates

SC '01 Paper Acceptance Rate 60 of 240 submissions, 25%;
Overall Acceptance Rate 1,516 of 6,373 submissions, 24%

Upcoming Conference

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)3
  • Downloads (Last 6 weeks)0
Reflects downloads up to 13 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2018)Optimizing MPI collective communication by orthogonal structuresCluster Computing10.1007/s10586-006-9740-99:3(257-279)Online publication date: 24-Dec-2018
  • (2004)Improving the execution time of global communication operationsProceedings of the 1st conference on Computing frontiers10.1145/977091.977130(276-287)Online publication date: 14-Apr-2004
  • (2004)A source code analyzer for performance prediction18th International Parallel and Distributed Processing Symposium, 2004. Proceedings.10.1109/IPDPS.2004.1303333(262-269)Online publication date: 2004
  • (2002)Library support for hierarchical multi-processor tasksProceedings of the 2002 ACM/IEEE conference on Supercomputing10.5555/762761.762782(1-10)Online publication date: 16-Nov-2002

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media