Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

The cluster coffer: : Teaching HPC on the road

Published: 01 September 2021 Publication History

Highlights

Showing the feasibility of constructing small, portable HPC systems for education.
Illustrating the benefits of this system for live interaction and application steering.
Providing enough material and information for others to reproduce and extend our work.

Abstract

Teaching parallel programming and HPC is a difficult task. There is a large number of sophisticated hardware and software components, each complex on their own and often showing non-intuitive interaction when used in combination. We consider education in HPC among the more difficult topics in computer science due to the fact that larger distributed memory systems are ubiquitous yet inaccessible and intangible to students. In this work, we present the Cluster Coffer, a miniature cluster computer based on 16 ARM compute boards that we believe is suitable for reducing the entry barrier to HPC in teaching and public outreach. We discuss our design goals for providing a portable, inexpensive system that is easy to maintain and repair. We outline the implementation path we took in terms of hardware and software, in order to provide others with the information required to reproduce and extend our work. Finally, we present two use cases for which the Cluster Coffer has been used multiple times, and will continue to be used in the upcoming years.

References

[1]
A. Auweter, A. Bode, M. Brehm, L. Brochard, N. Hammer, H. Huber, R. Panda, F. Thomas, T. Wilde, A case study of energy aware scheduling on SuperMUC, in: J.M. Kunkel, T. Ludwig, H.W. Meuer (Eds.), Supercomputing, Springer International Publishing, Cham, 2014, pp. 394–409.
[2]
D. Bedard, M.Y. Lim, R. Fowler, A. Porterfield, PowerMon: fine-grained and integrated power monitoring for commodity computer systems, in: Proceedings of the IEEE SoutheastCon 2010, SoutheastCon, 2010, pp. 479–484,.
[3]
Y. Censor, Pareto optimality in multiobjective problems, Discrete Appl. Math. 4 (1) (1977) 41–59.
[4]
S.J. Cox, J.T. Cox, R.P. Boardman, S.J. Johnston, M. Scott, N.S. O'brien, Iridis-Pi: a low-cost, compact demonstration cluster, Clust. Comput. 17 (2) (2014) 349–358.
[5]
Dongarra, J.; Luszczek, P.; Heroux, M.A.; Ye, K. (2021): HPCG benchmark. https://www.hpcg-benchmark.org/.
[7]
Engineering, D.S. (2020): Das U-Boot – The Universal Boot Loader. https://www.denx.de/wiki/U-Boot.
[8]
V.W. Freeh, N. Kappiah, D.K. Lowenthal, T.K. Bletsch, Just-in-time dynamic voltage scaling: exploiting inter-node slack to save energy in MPI programs, J. Parallel Distrib. Comput. 68 (9) (2008) 1175–1185.
[9]
S.J. Johnston, P.J. Basford, C.S. Perkins, H. Herry, F.P. Tso, D. Pezaros, R.D. Mullins, E. Yoneki, S.J. Cox, J. Singer, Commodity single board computer clusters and their applications, Future Gener. Comput. Syst. 89 (2018) 201–212,. http://www.sciencedirect.com/science/article/pii/S0167739X18301833.
[10]
H. Jordan, T. Heller, P. Gschwandtner, P. Zangerl, P. Thoman, D. Fey, T. Fahringer, The allscale runtime application model, in: 2018 IEEE International Conference on Cluster Computing, CLUSTER, 2018, pp. 445–455,.
[11]
H. Jordan, P. Gschwandtner, P. Thoman, P. Zangerl, A. Hirsch, T. Fahringer, T. Heller, D. Fey, The allscale framework architecture, Parallel Comput. (2020),. http://www.sciencedirect.com/science/article/pii/S0167819120300417.
[12]
J.H. Laros, P. Pokorny, D. DeBonis, PowerInsight - a commodity power measurement capability, in: 2013 International Green Computing Conference Proceedings, 2013, pp. 1–6,.
[13]
Ltd, M.P. (2020): Scalable clusters make HPC R&D easy as Raspberry Pi. http://bitscope.com/blog/FM/?p=GF13L.
[14]
Marinos, A. (2014): What would you do with a 120-Raspberry Pi Cluster?. https://www.balena.io/blog/what-would-you-do-with-a-120-raspberry-pi-cluster/.
[15]
S. Markidis, G. Lapenta Rizwan-uddin, Multi-scale simulations of plasma with iPIC3D, Math. Comput. Simul. 80 (7) (2010) 1509–1519,. multiscale modeling of moving interfaces in materials http://www.sciencedirect.com/science/article/pii/S0378475409002444.
[16]
F. Nielsen, Introduction to HPC With MPI for Data Science, 2016.
[17]
OpenMP Architecture Review Board (May 2018): OpenMP application program interface version 5.0. https://www.openmp.org/spec-html/5.0/openmp.html.
[18]
Petitet, A.; Whaley, R.C.; Dongarra, J.; Cleary, A. (2018): HPL - a portable implementation of the high-performance linpack benchmark for distributed-memory computers. https://www.netlib.org/benchmark/hpl/.
[19]
A.M. Pfalzgraf, J.A. Driscoll, A low-cost computer cluster for high-performance computing education, in: IEEE International Conference on Electro/Information Technology, IEEE, 2014, pp. 362–366.
[21]
T.R. Scogland, C.P. Steffen, T. Wilde, F. Parent, S. Coghlan, N. Bates, W.-c. Feng, E. Strohmaier, A power-measurement methodology for large-scale, high-performance computing, in: Proceedings of the 5th ACM/SPEC International Conference on Performance Engineering, ICPE '14, Association for Computing Machinery, New York, NY, USA, 2014, pp. 149–159,.
[22]
Strohmaier, E.; Dongarra, J.; Simon, H.; Meuer, M. (1993): CM-5/1024. https://www.top500.org/system/166997/.
[23]
Strohmaier, E.; Dongarra, J.; Simon, H.; Meuer, M. (1994): XP/S140. https://www.top500.org/system/172538/.
[24]
Strohmaier, E.; Dongarra, J.; Simon, H.; Meuer, M. (2020): Graham. https://www.top500.org/system/179047/.
[25]
Khronos OpenCL Working Group, SYCL Specification 1.2.1, Tech. rep., Khronos OpenCL Working Group 2017.
[26]
Top500 (2020): Supercomputer Fugaku. https://www.top500.org/system/179807/.
[27]
F.P. Tso, D.R. White, S. Jouet, J. Singer, D.P. Pezaros, The Glasgow Raspberry Pi Cloud: a scale model for cloud computing infrastructures, in: 2013 IEEE 33rd International Conference on Distributed Computing Systems Workshops, 2013, pp. 108–112,.
[28]
Wikipedia (2020): March 1989 geomagnetic storm. https://en.wikipedia.org/wiki/March_1989_geomagnetic_storm.
[29]
S. Williams, A. Waterman, D. Patterson, Roofline: an insightful visual performance model for multicore architectures, Commun. ACM 52 (4) (2009) 65–76,.

Cited By

View all
  • (2022)Optimized Page Fault Handling During RDMAIEEE Transactions on Parallel and Distributed Systems10.1109/TPDS.2022.317566633:12(3990-4005)Online publication date: 1-Dec-2022

Index Terms

  1. The cluster coffer: Teaching HPC on the road
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Information & Contributors

        Information

        Published In

        cover image Journal of Parallel and Distributed Computing
        Journal of Parallel and Distributed Computing  Volume 155, Issue C
        Sep 2021
        121 pages

        Publisher

        Academic Press, Inc.

        United States

        Publication History

        Published: 01 September 2021

        Author Tags

        1. 00-01
        2. 99-00

        Author Tags

        1. High performance computing
        2. Teaching
        3. Parallel programming
        4. Portability
        5. Public outreach

        Qualifiers

        • Research-article

        Contributors

        Other Metrics

        Bibliometrics & Citations

        Bibliometrics

        Article Metrics

        • Downloads (Last 12 months)0
        • Downloads (Last 6 weeks)0
        Reflects downloads up to 11 Feb 2025

        Other Metrics

        Citations

        Cited By

        View all
        • (2022)Optimized Page Fault Handling During RDMAIEEE Transactions on Parallel and Distributed Systems10.1109/TPDS.2022.317566633:12(3990-4005)Online publication date: 1-Dec-2022

        View Options

        View options

        Figures

        Tables

        Media

        Share

        Share

        Share this Publication link

        Share on social media