Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3176364.3176365acmotherconferencesArticle/Chapter ViewAbstractPublication PageshpcasiaConference Proceedingsconference-collections
research-article
Open access

Performance evaluation for a hydrodynamics application in XcalableACC PGAS language for accelerated clusters

Published: 31 January 2018 Publication History

Abstract

Clusters equipped with accelerators such as GPUs and MICs are used widely. To use these clusters, programmers write programs for their applications by combining MPI with one of the accelerator programming models such as CUDA and OpenACC. The accelerator programming component is becoming easier because of a directive-based OpenACC, but complex distributed-memory programming using MPI means that programming is still difficult. In order to simplify the programming process, XcalableACC (XACC) has been proposed as an "orthogonal" integration of the PGAS language XcalableMP (XMP) and OpenACC. XACC provides the original XMP and OpenACC features, as well as their extensions for communication between accelerator memories. In this study, we implemented a hydrodynamics mini-application Clover-Leaf in XACC and evaluated the usability of XACC in terms of it performance and productivity. According to the performance evaluation, the XACC version achieved 87--95% of the performance of the MPI+CUDA version and 93--101% of the MPI+OpenACC version with strong scaling, and 88--91% of the MPI+CUDA version and 94--97% of the MPI+OpenACC version with weak scaling. In particular, the halo exchange time was better with XACC than MPI+OpenACC in some cases because the Omni XACC runtime is written in MPI and CUDA, and it is well tuned. The productivity evaluation showed that the application could be implemented after small changes compared with the serial version. These results demonstrate that XACC is a practical programming language for science applications.

References

[1]
Center for Computational Sciences, University of Tsukuba. 2017. Supercomputers - Center for Computational Sciences. (2017). Retrieved August 11, 2017 from https://www.ccs.tsukuba.ac.jp/eng/supercomputers/#HA-PACS
[2]
T. Hanawa, Y. Kodama, T. Boku, and M. Sato. 2013. Interconnection Network for Tightly Coupled Accelerators Architecture. In High-Performance Interconnects (HOT), 2013 IEEE 21st Annual Symposium on. 79--82.
[3]
Francis H. Harlow and J. Eddie Welch. 1965. Numerical Calculation of Time-Dependent Viscous Incompressible Flow of Fluid with Free Surface. The Physics of Fluids 8, 12 (1965), 2182--2189.
[4]
J. A. Herdman, W. P. Gaudin, S. McIntosh-Smith, M. Boulton, D. A. Beckingsale, A. C. Mallinson, and S. A. Jarvis. 2012. Accelerating Hydrocodes with OpenACC, OpenCL and CUDA. In 2012 SC Companion: High Performance Computing, Networking Storage and Analysis. 465--471.
[5]
Charles H. Koelbel and Mary E. Zosel. 1993. The High Performance FORTRAN Handbook. MIT Press, Cambridge, MA, USA.
[6]
A. C. Mallinson, David A. Beckingsale, W. P. Gaudin, J. A. Herdman, and Stephen A. Jarvis. 2013. Towards Portable Performance for Explicit Hydrodynamics Codes. In 1st International Workshop on OpenCL (IWOCL 13).
[7]
A. C. Mallinson, S. A. Jarvis, W. P. Gaudin, and J. A. Herdman. 2014. Experiences at Scale with PGAS Versions of a Hydrodynamics Application. In Proceedings of the 8th International Conference on Partitioned Global Address Space Programming Models (PGAS '14). ACM, New York, NY, USA, Article 9, 11 pages.
[8]
Mantevo.org. 2017. Home of the Mantevo Project. (2017). Retrieved August 11, 2017 from https://mantevo.org/
[9]
Masahiro Nakao, Hitoshi Murai, Hidetoshi Iwashita, Akihiro Tabuchi, Taisuke Boku, and Mitsuhisa Sato. 2017. Productivity and Performance of XcalableACC Language for Acceralated Clusters. (Feb 2017). Poster presented at The 7th AICS International Symposium.
[10]
Masahiro Nakao, Hitoshi Murai, Takenori Shimosaka, Akihiro Tabuchi, Toshihiro Hanawa, Yuetsu Kodama, Taisuke Boku, and Mitsuhisa Sato. 2014. XcalableACC: Extension of XcalableMP PGAS Language Using OpenACC for Accelerator Clusters. In Proceedings of the First Workshop on Accelerator Programming Using Directives (WACCPD '14). 27--36.
[11]
OpenACC-standard.org. 2017. OpenACC. (2017). Retrieved August 11, 2017 from http://www.openacc.org
[12]
PC Cluster Consortium. 2017. PC Cluster Consortium. (2017). Retrieved August 11, 2017 from http://www.pccluster.org
[13]
John Reid. 2010. Coarrays in the next Fortran Standard. ISO/IEC JTC1/SC22/WG5N1824 (Apr. 2010).
[14]
RIKEN AICS and University of Tsukuba. 2017. Omni Compiler Project. (2017). Retrieved August 11, 2017 from http://omni-compiler.org
[15]
RIKEN AICS and University of Tsukuba. 2017. XcalableACC Launguage Specification Version 1.0. (2017). Retrieved August 11, 2017 from http://xcalablemp.org/download/XACC/xacc-spec-1.0.pdf
[16]
Akihiro Tabuchi, Masahiro Nakao, Hitoshi Murai, Taisuke Boku, and Mitsuhisa Sato. 2017. Implementation and Evaluation of One-sided PGAS Communication in XcalableACC for Accelerated Clusters. In Proceedings of the 17th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGrid '17). 625--634.
[17]
Keisuke Tsugane, Taisuke Boku, Hitoshi Murai, Mitsuhisa Sato, William Tang, and Bei Wang. 2016. Hybrid-view programming of nuclear fusion simulation code in the PGAS parallel programming language XcalableMP. Parallel Comput. 57 (2016), 37--51.
[18]
UK Mini-App Consortium. 2017. UK Mini-App Consortium. (2017). Retrieved October 27, 2017 from http://uk-mac.github.io/
[19]
XcalableMP Specification Working Group. 2017. XcalableMP WebSite. (2017). Retrieved August 11, 2017 from http://www.xcalablemp.org

Cited By

View all
  • (2018)Evaluating the Impact of Proposed OpenMP 5.0 Features on Performance, Portability and Productivity2018 IEEE/ACM International Workshop on Performance, Portability and Productivity in HPC (P3HPC)10.1109/P3HPC.2018.00007(37-46)Online publication date: Nov-2018

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
HPCAsia '18 Workshops: Proceedings of Workshops of HPC Asia
January 2018
86 pages
ISBN:9781450363471
DOI:10.1145/3176364
This work is licensed under a Creative Commons Attribution International 4.0 License.

Sponsors

  • IPSJ: Information Processing Society of Japan

In-Cooperation

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 31 January 2018

Check for updates

Author Tags

  1. GPU
  2. PGAS
  3. XcalableMP
  4. accelerator
  5. cluster
  6. hydrodynamics
  7. openACC

Qualifiers

  • Research-article

Conference

HPC Asia 2018 WS
Sponsor:
  • IPSJ
HPC Asia 2018 WS: Workshops of HPC Asia 2018
January 31, 2018
Tokyo, Chiyoda

Acceptance Rates

Overall Acceptance Rate 69 of 143 submissions, 48%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)45
  • Downloads (Last 6 weeks)11
Reflects downloads up to 06 Oct 2024

Other Metrics

Citations

Cited By

View all
  • (2018)Evaluating the Impact of Proposed OpenMP 5.0 Features on Performance, Portability and Productivity2018 IEEE/ACM International Workshop on Performance, Portability and Productivity in HPC (P3HPC)10.1109/P3HPC.2018.00007(37-46)Online publication date: Nov-2018

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media