Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
An Efficient Privacy and Anonymity Setup on Hyperledger Fabric for Blockchain-Enabled Internet of Things (IoT) Devices
Previous Article in Journal
Object Detection, Segmentation and Categorization in Artificial Intelligence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evolving High-Performance Computing Data Centers with Kubernetes, Performance Analysis, and Dynamic Workload Placement Based on Machine Learning Scheduling

1
Department of Operating Systems, Algebra University, 10000 Zagreb, Croatia
2
Department of Control and Computer Engineering, Faculty of Electrical Engineering and Computing, University of Zagreb, 10000 Zagreb, Croatia
*
Authors to whom correspondence should be addressed.
Electronics 2024, 13(13), 2651; https://doi.org/10.3390/electronics13132651
Submission received: 31 May 2024 / Revised: 28 June 2024 / Accepted: 2 July 2024 / Published: 5 July 2024
(This article belongs to the Section Computer Science & Engineering)

Abstract

:
In the past twenty years, the IT industry has moved away from using physical servers for workload management to workloads consolidated via virtualization and, in the next iteration, further consolidated into containers. Later, container workloads based on Docker and Podman were orchestrated via Kubernetes or OpenShift. On the other hand, high-performance computing (HPC) environments have been lagging in this process, as much work is still needed to figure out how to apply containerization platforms for HPC. Containers have many advantages, as they tend to have less overhead while providing flexibility, modularity, and maintenance benefits. This makes them well-suited for tasks requiring a lot of computing power that are latency- or bandwidth-sensitive. But they are complex to manage, and many daily operations are based on command-line procedures that take years to master. This paper proposes a different architecture based on seamless hardware integration and a user-friendly UI (User Interface). It also offers dynamic workload placement based on real-time performance analysis and prediction and Machine Learning-based scheduling. This solves a prevalent issue in Kubernetes: the suboptimal placement of workloads without needing individual workload schedulers, as they are challenging to write and require much time to debug and test properly. It also enables us to focus on one of the key HPC issues—energy efficiency. Furthermore, the application we developed that implements this architecture helps with the Kubernetes installation process, which is fully automated, no matter which hardware platform we use—x86, ARM, and soon, RISC-V. The results we achieved using this architecture and application are very promising in two areas—the speed of workload scheduling and workload placement on a correct node. This also enables us to focus on one of the key HPC issues—energy efficiency.

1. Introduction

The shift from conventional data center designs to more dynamic and efficient models is becoming increasingly critical in modern IT. Traditional data centers, which typically rely on fixed resource allocation and design patterns, must be better suited to meet the variable demands of contemporary workloads, especially when dealing with HPC (high-performance computing). HPC traditionally relies on specialized hardware and software to achieve maximum performance. The integration of container orchestration platforms, such as Kubernetes, has opened new avenues for deploying and managing HPC workloads more efficiently. Studies have shown that CPU hardware-assisted hypervisor features can significantly impact performance [1]. That does not mean virtual machines (VMs) do not have inherent advantages—for example, live migration. Multiple VM live migration scheduling has been explored to optimize VMware vMotion efficiency [2]. This is not something that we can achieve on physical servers or containers.
Regarding efficiency, virtual machines can be scheduled to be placed on optimal nodes using concepts like VMware DRS (Distributed Resource Scheduler). DRS has been compared to human experts, highlighting the benefits of automated systems [3]. Over the years, its ability to act less robotically and more environmentally aware has also improved, especially with features like predictive DRS. Performance overhead comparisons between hypervisors and containers indicate that containers offer superior efficiency [4]. We want to use this to our advantage in HPC as efficiency and overhead are areas where underlying technology for HPC applications can improve significantly.
Deploying Docker containers on bare metal and virtual machines has significantly improved performance, offering a scalable, high-performance alternative to traditional hypervisors. When integrated with Kubernetes and DevOps workflows, containers become more accessible. While not without its complexities in design, deployment, correct configuration, and lifecycle management, this transition is proving its worth in large organizations. The practical benefits of container orchestration in these settings are a reassuring sign that the transition is worth it, as containers emerge as a viable, high-performance alternative to hypervisors.
Previous research shows that virtualization (KVM) and containerization have overheads [5]. For example, KVM has a bit more overhead for every I/O operation, which influences the overall performance depending on the I/O size [5]. This is why containers are mostly a better choice for I/O-sensitive workloads. Kubernetes also helps in that regard, as it has become a powerful tool for managing large-scale containerized applications, inspiring us about future possibilities [6]. Studies comparing Docker and Podman show significant performance benefits [7].
Regarding containers and their application in HPC, rootless containers with Podman have been evaluated, providing additional security and flexibility [8]. Using containerization in scientific calculations underscores its suitability for HPC workloads [9]. Comparisons of container systems for Machine Learning scenarios show Docker and Podman as effective solutions [10]. While ramping up our research, these efforts helped us conduct more research on how Kubernetes interacts with workload scheduling. This part of our research ended up with a firm realization of how unsuited the default Kubernetes scheduler is for HPC applications. This is why fine-grained scheduling for containerized HPC workloads in Kubernetes clusters can ensure efficient resource management [11]. We should also take note of the practical benefits of using containers—they offer reassurance about the transition, as there are clear advantages to using Kubernetes with HPC workloads [12].
Research is slowly evaluating the efficiency and resource allocation of Kubernetes with HPC [13,14,15,16,17]. Some research is also being carried out on implementing QoS (Quality of Service) principles in Kubernetes/HPC scenarios, as autoscaling pods on on-premises Kubernetes infrastructure with QoS awareness can support consistent performance [18]. This influences how we package HPC applications, as OpenMP (Open Multi-Processing) and MPI (Message Passing Interface) have specific requirements to be met for them to operate correctly. Design considerations for containerized MPI applications highlight the complexities of HPC workload management [19]. In terms of monitoring, there are areas for which Kubernetes does not have a native, built-in solution, such as hardware monitoring. This is where ML techniques for detecting cluster anomalies support proactive system management [20].
Dynamic workload placement, guided by real-time metrics, resource availability, and energy efficiency, enables significant operational efficiency and sustainability improvements. This approach provides a more agile and responsive infrastructure that can adapt to diverse workload requirements, ensuring optimal performance while reducing energy usage and costs.
Implementing such dynamic strategies within data center operations marks a significant advancement, promising better resource utilization, a reduced carbon footprint, and alignment with modern business needs. This paper examines the technologies and strategies driving this transformation, focusing on the pivotal role of hardware and software integration orchestrated by Kubernetes in fostering a more efficient and sustainable HPC infrastructure. As we researched this topic, we built a fully functional platform that can do everything stated in this paper—install or add nodes to the Kubernetes cluster, perform a dynamic evaluation, manage Kubernetes clusters, and place workloads based on Machine Learning (ML) input, as we will show later in this paper.
This paper aims to propose a new software platform to apply containers and Kubernetes to HPC efficiently. The hardware part of that architecture can be designed as it has always been. The big difference between our approach and the traditional approach is how hardware and software interact in scheduling HPC workloads. Our ML-driven (Machine Learning) approach to workload placement offers many advantages, especially when integrated with an easy-to-use UI (User Interface).
This paper is organized as follows. The next sections provide relevant information about all technological aspects of the paper—containerization, container orchestration, virtualization technologies, and a general overview of HPC. Then, we will discuss real-life problems when applying Kubernetes to HPC workloads and offer our research and insight into how to solve these problems. This necessitated the creation of an entirely new software platform that aims to solve these problems in a way that is user-friendly and as automated as possible. One of our primary goals was to simplify how we create container workloads for our HPC environment and then schedule these workloads to be placed on the best possible node as calculated and recommended by a built-in AI engine. In our software platform, we can also override these recommendations and manually place workloads if we wish to do so. We described the way we designed our application in a separate section. The last sections of our paper are reserved for preliminary results, a discussion about future research directions, limitations, and our conclusions.

2. Technology Overview

Virtualization, containers, orchestrators, and high-performance computing (HPC) are fundamental technologies in today’s IT landscape, each enhancing the efficiency and scalability of computing environments in unique ways. Virtualization enables the creation of multiple virtual machines (VMs) from a single physical hardware resource, improving resource utilization and isolation. Containers, managed by tools like Docker and Podman, provide a lightweight alternative to VMs by encapsulating applications and their dependencies in isolated environments, thus ensuring portability and consistency across different deployment platforms. Orchestrators such as Kubernetes and OpenShift automate containerized applications’ deployment, scaling, management, and high availability and streamline operations in complex cloud infrastructures. HPC utilizes advanced architectures and parallel processing to handle demanding computational tasks, facilitating significant advancements in scientific research, engineering, and large-scale data analysis. Collectively, these technologies form the backbone of the modern digital ecosystem, driving innovation and enhancing operational efficiency across various industries. In this section, we will go through an overview of all the technologies related to our paper.

2.1. Virtualization

Virtualization abstracts physical hardware to create multiple simulated environments called virtual machines (VMs) from a single physical system by creating a virtual framework that uses code that mimics physical equipment functionality [21]. Each VM operates its operating system (OS) and applications, isolated from other VMs, using a hypervisor. Hypervisors are categorized into Type 1 (bare metal) and Type 2 (hosted).
Type 1 Hypervisors are installed directly on physical hardware, acting as an intermediary layer between the hardware and the VMs. Examples include VMware ESXi, Microsoft Hyper-V, and Xen. They are more efficient and perform much better since bare-metal hypervisors operate at the most privileged CPU level [1].
Type 2 or hosted hypervisors are executed on the host operating system as a typical application [1]. The most prominent examples include Oracle VirtualBox and VMware Workstation. These hypervisors are typically easier to set up and are used primarily for development and testing.
A primary advantage of virtualization is the strong isolation it offers. Each VM runs in a self-contained environment, enhancing security and stability. If one VM fails or is compromised, the others remain unaffected.
However, this strong isolation comes at the cost of overhead. Each VM includes a full OS, leading to higher memory and storage consumption. This can make virtualization less efficient regarding resource usage than containerization.
Another significant advantage of virtualization is its support for live migration, which allows a running virtual machine to move to another physical host without disrupting its service [2].
Virtualization facilitates better resource allocation and load balancing. Virtual machines can dynamically allocate CPUs, memory, and storage resources based on current needs, ensuring optimal performance and efficiency. Various policy-based approaches exist, such as resource pooling, reservations, limits, etc. Tools like VMware’s Distributed Resource Scheduler (DRS) can also automatically balance workloads across multiple hosts, enhancing resource utilization. DRS takes resource management decisions according to metrics related to VMs, hosts, and clusters for memory and CPU (network and storage not considered) [3].

2.2. Containerization Technology

Containerization technologies have transformed application deployment and management by offering lightweight, portable, and consistent environments. It has been a component of the Linux kernel since 2008 [22]. This innovation began with BSD jails, which introduced isolated user spaces. Docker later popularized containerization by streamlining container creation and management. Linux Containers (LXC) advanced the field by providing OS-level virtualization for Linux. Recently, Podman has emerged as a Docker-compatible alternative, offering daemonless container management and enhancing security and flexibility in container operations.
Containers offer significant advantages when compared to virtual machines. Some of the most important benefits include the following:
  • Efficient resource usage: Containers share the host system’s kernel, eliminating the need for a complete operating system for each instance. This results in lower overhead than virtual machines (VMs), each requiring their own OS, leading to more efficient CPU and memory use. The container’s average performance is generally better than that of the VM and is comparable to that of the physical machine in terms of many features [4].
  • Faster startup time: Containers can start almost instantly because they do not require booting an entire operating system. In contrast, VMs take significantly longer to start as they need to initialize a whole OS, and the startup time difference from server power-on can be up to 50% [23]. This speed advantage makes containers ideal for applications that require quick scaling and deployment.
  • Density: Due to their lightweight nature, a single host can run many more containers than VMs. For certain types of workloads, an application container’s startup time is 16× lower than that of a VM, and its memory footprint is 60× lower than that of a VM [24]. This higher density allows for better utilization of hardware resources, enabling more applications to run on the same infrastructure.
  • Consistency: Containers encapsulate the application and its dependencies, ensuring consistent behavior across different environments. This consistency across development, testing, and production reduces bugs related to environmental differences.
  • Better dependency management: Containers package all the necessary dependencies for an application, eliminating conflicts that arise when applications require different versions of the same dependencies. This encapsulation simplifies dependency management and ensures reliable application performance across various environments.
Let us delve into some of the architectural details of these containerization methodologies.

2.3. BSD Jails

BSD jails are an efficient and secure virtualization mechanism in FreeBSD, introduced with FreeBSD 4.0 in 2000. They allow administrators to partition a FreeBSD system into multiple isolated mini-systems called jails. In BSD jails, the operating system should be the point of most significant gain since the trapped systems are based on sharing the running kernel with the base system, thus eliminating the need for additional resources related to implementing a new system instance operating [25].
The jails can be assigned unique IP addresses and configured with specific network settings, enabling different software versions to run simultaneously. This makes them particularly advantageous in testing and development settings and allows the safe testing of updates or new configurations without risking the host system’s stability.

2.4. Docker

Docker, a platform for containerization, revolutionizes virtualization by packaging applications and their dependencies into containers, ensuring uniformity across different environments. This method contrasts sharply with traditional virtualization, where each virtual machine (VM) contains a complete guest operating system, leading to significant resource and performance overhead. Docker employs containerization to isolate applications from the host system, utilizing Linux kernel features such as namespaces and control groups (cgroups) to create lightweight containers.
As we can see in Figure 1, the Docker architecture comprises several vital components. The Docker Engine, the core of Docker, includes a daemon (dockerd) that manages container lifecycles, images, networks, and storage. This daemon interacts with the host’s kernel to create and manage containers. The Docker Client is a command-line interface (CLI) used to interact with the Docker daemon, providing commands to build, run, and manage containers. Docker Hub, a cloud-based registry service, hosts Docker images, facilitating easy sharing and distribution.
Docker containers share the host OS kernel, resulting in lower overhead than VMs. This shared kernel model allows for higher density and efficiency, as containers require fewer resources and start faster than VMs. A Felter et al. (2015) study demonstrated that Docker containers achieve near-native performance for CPU, memory, and I/O operations, often outperforming traditional VMs.
Managing many containers in production environments necessitates orchestration tools. Kubernetes is the most prominent tool for orchestrating Docker containers, providing automated deployment, scaling, and management of containerized applications. Kubernetes manages container lifecycles and automates their deployment, scaling, and management. With additional configuration, it can also manage network access, security, load balancing, and resource allocation across a cluster of nodes, providing high availability. Docker transforms the deployment and management of applications through efficient containerization, leveraging kernel features to offer lightweight virtualization. Security considerations must be addressed to ensure robust isolation despite its significant performance and resource utilization benefits. This is why it is necessary to protect the accuracy of the data, and the data’s integrity must always be preserved [26]. Integrating orchestration tools like Kubernetes further enhances Docker’s capabilities, making it an indispensable technology in modern DevOps practices. The term “DevOps” is a combination of the terms “development” and “operations” [27], which refers to the unification of teams, transforming the existing silos into a set of teams that focus work on the organization and not on the activity within it, thus linking all the steps that already exist in software development (including delivery) [28]. For these mechanisms to work, the Dev and Ops teams must collaborate and support each other [29]. Kubernetes is a core component of DevOps implementation.
Although Docker as a project did not invent containerization, it brought containers to the foreground in IT. It made them what they are today—the most prominent technology to package applications worldwide. The concept was not all that new—in UNIX systems; we could use chroot environments that could be traced back to the beginning of the containerization idea at the end of the 1970s. We might consider the chroot idea the first idea that very crudely separated processes from the disk space, and out of that idea, in the 2000s, BSD jails were introduced.

2.5. LXC

After BSD jails, the LXC (Linux Containers) project was introduced to the market in 2008. Linux Containers (LXCs) offer a lightweight virtualization method that provides an environment like a complete Linux system but without the overhead of a separate kernel. LXCs utilize kernel features such as IPC, UTS, mount, PID, network, and user namespaces to isolate processes. These namespaces ensure that processes within a container cannot interact with those outside, enhancing security.
One of the significant advantages of LXC is its simplicity, flexibility, and low overhead. Unlike traditional virtual machines, which require separate kernels and virtualized hardware, LXCs share the host system’s kernel, making them more efficient regarding resource usage [30]. LXC also offers comprehensive tools and templates for creating and managing containers. For instance, users can utilize lxc-create to create a container and lxc-start to initiate it. Containers can be configured to autostart, and their configurations can be managed through various commands and configuration files [31].
LXC’s ability to fine-tune resource allocation using cgroups further enhances its efficiency, enabling precise control over CPU, memory, and I/O resources. As a result, LXCs are widely used in enterprise and cloud computing environments to optimize performance and reduce costs [5,6].

2.6. Podman

Podman, short for “Pod Manager”, is an open-source tool designed for developing, managing, and running Linux container systems. Developed by Red Hat engineers, Podman is part of the libpod library and is notable for its daemonless architecture, which enhances security and reduces overhead compared to Docker. Unlike Docker, Podman does not require a continuous background service (daemon) to manage containers. This makes it a more secure option by allowing containers to run without root privileges and thus minimizing potential security risks associated with root-owned processes [7].
Podman supports Open Container Initiative (OCI) standards and ensures compatibility with various container images and runtimes. Its command-line interface (CLI) is like Docker’s, facilitating an easier transition for users. Podman integrates with other container management tools, such as Buildah for image building and Skopeo for image transfer between different registries [8].
Podman also manages pods, groups of containers that share the same resources, like Kubernetes pods, as shown in Figure 2. This feature benefits complex applications requiring multiple containers to work together seamlessly. Additionally, Podman offers a RESTful API for remote management and supports various Linux distributions, including Fedora, CentOS, and Ubuntu [9].
The introduction of Podman in 2019 marked a shift away from Docker’s dominance in the market. Podman addressed a significant issue with Docker: its monolithic architecture, which was seen as overly complex for large-scale designs, particularly concerning the high availability of the Kubernetes control plane [10]. While these platforms were not initially designed for high-performance computing (HPC) environments and large heterogeneous clusters, the idea has gradually gained traction. It is now becoming mainstream in the HPC world. HPC is an ideal use case for applications running close to the hardware to avoid costly indirection layers such as virtualization [32] or CI/CD pipelines [33].

2.7. Kubernetes

Docker and Podman’s popularity led to the development of the next set of products for efficient management and scaling (Kubernetes and OpenShift). We are deliberately skipping the idea of Docker Swarm as it is obsolete and not applicable to standard, let alone HPC use cases [34]. These two products (Kubernetes and OpenShift) are, in essence, container management and orchestration platforms. Container orchestration technology was initially developed and applied inside large IT companies like Google and Yandex to deploy scalable high-load services [35]. Figure 3 illustrates the standard Kubernetes architecture orchestrating containers:
Kubernetes is an open-source platform for container orchestration. It is essential in today’s cloud computing landscape because it can automate containerized applications’ deployment, scaling, and management. Its architecture offers a robust, scalable, and flexible management system for containerized applications across a cluster of nodes.
Kubernetes employs primary-worker architecture at its foundation. The primary node is pivotal in maintaining the cluster’s desired state, orchestrating tasks, and managing the overall lifecycle of the cluster. Key components of the primary node include the API Server, which acts as the control plane’s front end, handling all RESTful API requests. The etcd service is a key-value store that persists cluster state and configuration data. The Controller Manager manages controllers for routine tasks like node health monitoring and replication and allows the cluster to be linked to the cloud platform [36]. Lastly, the scheduler assigns workloads to specific nodes based on resource availability and predefined constraints [37].
Worker nodes, which host the containerized applications, include several critical components. The Kubelet is an agent running on each node, ensuring containers within a Pod run. It communicates with the primary node to receive commands and reports the node’s status. The Kube-Proxy manages network communication within the cluster, implementing network rules for Pod connectivity. The Container Runtime (e.g., Docker, containerd) is also responsible for running the containers [38].
Kubernetes organizes applications into Pods, the minor deployable units containing one or more containers. Pods share the same network namespace, allowing communication via localhost. They are designed to be ephemeral, meaning they can be replaced or rescheduled as needed. Kubernetes uses ReplicaSets and deployments to manage Pods, ensuring the desired number of Pod replicas run at any time [39].
A key feature of Kubernetes is its ability to perform service discovery and load balancing. The Service abstraction provides a stable IP address and DNS name to a set of Pods, ensuring reliable access. Services can be exposed internally within the cluster and externally to the internet. Kubernetes Ingress Controller is used as the entry point for network traffic that directs traffic to the corresponding service based on rules [40].
Kubernetes offers Horizontal Pod Autoscaling (HPA) for scaling and resource management. HPA adjusts the number of Pod replicas based on the observed CPU utilization or other selected metrics. This ensures efficient resource utilization and allows the system to respond to workload demands dynamically [41].
Kubernetes emphasizes declarative configuration and automation. Users define the desired states using YAML or JSON manifests, and Kubernetes ensures that the actual state matches the desired state. This approach simplifies management and enhances the reproducibility and reliability of deployments [42]. It helps deliver a scalable, flexible, resilient platform for managing containerized applications. Its primary-worker model and robust abstractions for managing workloads and services make it a critical tool in modern cloud-native environments.
Now, if we expand our view of the problems we mentioned and add high availability of applications into the discussion, it becomes obvious why Kubernetes and OpenShift became so popular. Orchestrators reduce costs and allow mechanisms to be put in place that contribute to the application’s resilience, providing adaptiveness to different operation environments and even fault tolerance mechanisms [43]. They are a developer’s dream, as the hundreds or thousands of containers can run web applications to our heart’s content. We can easily load-balance access to them via Kubernetes/OpenShift load-balancing objects (Ingress controllers, various load-balancers like Traefik, etc.) and achieve high availability via workload balancing. Stateless application architectures like the ones usually used by Kubernetes and OpenShift are, therefore, much easier to code. Furthermore, if we create such applications, we do not have to deal with clustering complexity, as the concept of high availability via clustering becomes unnecessary. This is a perfect case for real-life Kubernetes and one of the most common design patterns.

2.8. OpenShift vs. Kubernetes

Kubernetes is an open-source container orchestration platform that automates deploying, scaling, and managing containerized applications. It uses a primary-worker architecture, where the primary node manages the cluster’s state, and worker nodes run the containers. The critical components of Kubernetes include the API Server, etcd, Controller Manager, Scheduler, and Kubelet [44]. Kubernetes is known for its modular and extensible nature, allowing integration with various plugins and tools to enhance functionality [45]. It offers powerful features for service discovery, load balancing, horizontal scaling, and declarative configuration management [46].
OpenShift, developed by Red Hat, is a Kubernetes distribution that extends Kubernetes with several enterprise-grade features and tools, making it easier to manage containerized applications. OpenShift includes all Kubernetes features and adds capabilities like integrated CI/CD (Continuous Integration/Continuous Deployment) pipelines, developer-friendly tools, enhanced security features, and multi-tenancy support [47]. Its architecture is designed to provide a comprehensive and secure platform for running containerized applications in enterprise settings [48].
One of the main differences between Kubernetes and OpenShift is the level of built-in support and integration. OpenShift includes a built-in image registry, integrated logging and monitoring tools, and a web-based management console, which is not provided by default in Kubernetes. This makes OpenShift a more turnkey solution that can also be used as a managed service in cloud environments [49], as we can see from Figure 4.
Moreover, OpenShift enforces stricter security policies by default. For instance, it includes Security Context Constraints (SCCs) that define a Pod’s actions, enhancing the platform’s overall security. While Kubernetes is highly customizable, achieving a similar security level requires more manual configuration [50].
Both Kubernetes and OpenShift offer robust platforms for managing containerized applications, but OpenShift provides a more comprehensive and enterprise-ready solution with additional features and integrated tools. Kubernetes remains a powerful option for users who prefer to customize their environment to specific needs. Our choice of orchestration platform for this paper will be Kubernetes, as we do not want to impose even more cost to our environment related to OpenShift licensing.

2.9. HPC

High-performance computing (HPC) is essential for solving complex computational problems across various scientific and engineering fields. These systems are built to provide maximum computational power, utilizing advanced architectures and innovative technologies to handle large datasets and execute extensive simulations efficiently.
HPC is a technology that configures multiple computing nodes into clusters to achieve high performance [51]. Modern HPC architectures typically involve multi-core processors, parallel computing techniques, and high-speed interconnects. Storage performance is a huge challenge for HPC storage systems, especially the ones with bandwidth-limited PFS (parallel file system) [52]. Advanced HPC systems use heterogeneous architectures, incorporating accelerators such as GPUs and FPGAs to improve performance and overall efficiency [53].
A crucial aspect of HPC architecture is the interconnect network, which links multiple compute nodes into a cohesive system. Efficient communication between nodes is vital for high performance, especially in massively parallel systems. Technology like InfiniBand HDR performs best for HPC, Cloud, and DL (Deep Learning) computing applications [54]. For years and years, all these moving parts have mainly been using virtualization.
HPC is applied in many fields requiring substantial computational resources. Everyday use cases include weather forecasting and climate modeling, molecular dynamics and bioinformatics, engineering and manufacturing, AI/ML training, oil and gas exploration, genomics, financial modeling, risk analysis, etc. [55,56,57,58,59].
HPC systems combine parallel computing techniques, multi-core processors, and specialized hardware accelerators to achieve high performance. Key components and architectural paradigms include:
  • Parallel Computing: HPC systems rely on parallel computing, where multiple processors perform computations simultaneously. This includes fine-grained parallelism, where tasks are divided into smaller subtasks, and coarse-grained parallelism, where larger independent tasks run concurrently [52].
  • Multi-Core and Many-Core Processors: These processors have multiple processing cores on a single chip and are connected with high-bandwidth communication links, including a QuickPath Interconnect (QPI) bus [53].
  • Accelerators and Heterogeneous Computing: HPC systems use accelerators like GPUs, FPGAs, and specialized processing units to handle specific calculations more efficiently than general-purpose CPUs. Heterogeneous computing combines CPUs with accelerators, leveraging the strengths of different processing units to optimize performance [54].
  • Distributed and Cluster Computing: HPC systems are often organized as clusters of interconnected computers (nodes), each with its own processors, memory, and storage. These clusters can scale to thousands of nodes, handling large datasets and complex simulations. Interconnects like InfiniBand and high-speed Ethernet enable fast node communication [57].
  • Memory Hierarchy and Storage: Due to the large data volumes processed, efficient memory management is critical in HPC systems. HPC architectures use multi-level memory hierarchies, including cache, main memory, and high-speed storage solutions, to ensure quick data access and minimize latency [58]. This is why we need a good strategy to port and optimize existing applications to a massively parallel processor (MPP) system [59].
Traditionally, HPC systems are typically equipped with workload managers that excel in specific tasks tailored to traditional use cases rather than new and emerging workloads such as AI. A workload manager comprises a resource manager and a job scheduler [60].
This concludes our technology overview section, which includes all technologies relevant to our paper. Let us discuss the differences between virtualization and containers to describe real-life problems when using containers for HPC workloads.

3. Virtualization vs. Containers

Virtualization and containerization are two fundamental technologies in modern computing, each with unique advantages and specific use cases. Virtualization involves creating multiple simulated environments (VMs) on a single physical system using hypervisors, which provides strong isolation but with higher overhead. A container engine enables lightweight and efficient application deployment using OS-level features to create isolated user spaces. Here are ten key differences between virtualization and Podman, used here as an example of the currently recommended container engine:
  • Architecture: Virtualization employs a hypervisor to create and manage VMs, each with its own OS. In contrast, Podman runs containers on a shared OS kernel, which reduces overhead and boosts efficiency.
  • Isolation: VMs achieve strong isolation by running separate OS instances. Podman containers provide process and filesystem isolation through namespaces and control groups, which, while less robust, are adequate for many applications.
  • Resource Overhead: Virtualization demands more resources because an entire OS instance is needed per VM. Podman containers are lightweight, sharing the host OS kernel and minimizing resource usage.
  • Performance: VMs generally have higher overhead and can be less efficient. Podman containers deliver better performance and faster startup times since they do not require a full OS boot.
  • Deployment Speed: Deploying VMs takes longer due to OS initialization. Podman containers can be deployed rapidly, making them ideal for quick development and scaling.
  • Resource Allocation: Resource allocation in virtualization is managed through the hypervisor and can be static or dynamic. Podman allows for flexible, real-time resource allocation for containers.
  • Use Cases: Virtualization is well suited for running diverse OS environments, legacy applications, and high-security workloads. Podman excels in microservices, CI/CD pipelines, and modern application development.
  • Security: VMs offer strong security with robust isolation, making them suitable for untrusted workloads. Podman containers are secure but rely on the host OS kernel, which can present vulnerabilities.
  • Maintenance: Managing VMs involves handling multiple OS instances, leading to higher maintenance overhead. Podman containers simplify maintenance by sharing the host OS and dependencies.
  • Compatibility: VMs can run different OS types and versions on a single host. Podman containers are limited to the host OS kernel but can quickly move across compatible environments.
Virtualization and containers have unique strengths and serve different purposes in modern IT infrastructure. Understanding these differences is crucial for selecting the appropriate technology for specific use cases and balancing isolation, performance, and resource efficiency. Therefore, there will be situations where using virtual machines will be much more convenient for us, even if that means more overhead and larger compute objects (virtual machines).
Let us now discuss the problems in terms of how to apply the idea of Kubernetes to HPC systems with workloads that can hugely benefit from hardware-specific acceleration.

4. Challenges Using Containers for HPC Workloads

HPC workloads have traditionally run on physical and virtual machines for many years and, for the most part, have stayed away from containers as a delivery mechanism. There are good reasons why virtualization technologies are often much more convenient than containerization technologies when discussing HPC use cases. One of the most important is how they allocate hardware. Advanced hardware capabilities such as PCIe (PCI Express) passthrough for dedicated hardware access from a virtual machine and SR-IOV (Single Root I/O Virtualization) for hardware partitioning are well established and familiar when using virtualization technologies while being sparsely available on containerization technologies. The primary reasons for this are not hardware-based but driver- and support-based. Previous research shows significant challenges in this area, as architectures like ConRDMA enable containers to use RDMA-allocated bandwidth more efficiently [61]. A lot of PCI Express cards we would like to use in HPC environments are not supported by container technologies. This might have a significant impact on how we accelerate HPC workloads, as we often need a GPU (Graphics Processing Unit), ASIC (Application-Specific Integrated Circuit), or FPGA (Field-Programmable Gate Array) to both accelerate workloads and make them much more efficient from compute/energy efficiency standpoint. However, HPC is a perfect use case for applications running as close to the hardware as possible to avoid costly indirection layers such as virtualization [12]. That is why we would like to use containers immediately for HPC workloads.
Containers are almost exclusively used to create and deploy modern web applications, not HPC applications. Typically, single-threaded applications (like most web applications) need a horizontal (scale-out) scaling approach, as the traditional way of vertical scaling (scale-up) does not increase performance. On the other hand, HPC applications are not designed to be run on a single core—they are often parallelized by forking as many threads as needed for the heavy data parallel portion [62]. Here are some common reasons for such a design:
  • HPC application problem size and complexity: HPC apps have large datasets and complex calculations that a single processor cannot efficiently manage. Because of this, the problem is divided into many small tasks that can be processed concurrently using multiple processors or nodes [13].
  • Speed and efficiency: Massively parallel design enables HPC apps to be quickly executed on HPC systems much faster than traditional computing methods. HPC applications can achieve significant speedups by utilizing hundreds, thousands, or more processors in parallel, solving problems much quicker than with a single processor [14].
  • Scalability: HPC application scalability is essential to HPC application design, especially as datasets grow larger and more complex problems are found, necessitating more computation power. Different parallel algorithms are developed to improve performance as we scale to more processors [15].
Multiple parallel computing models have been developed over the years to enable us to achieve these massively parallel speedups:
  • MPI (Message Passing Interface): Allows processors to communicate by sending and receiving MPI messages.
  • OpenMP (Open Multi-Processing): This enables us to perform parallel application programming for shared memory architecture.
  • CUDA (Compute Unified Device Architecture): Enables parallel processing on GPUs, often used for HPC because they are massively parallel hardware devices by design [24]. In GPU parallel programming and a CUDA framework, “kernels” and “threads” are foundational concepts for programming [63].
With this in mind, there are three significant operational problems when deciding to use containers running HPC workloads:
  • HPC applications are usually deployed as standard applications, meaning they must be re-packaged into layers and container images if we want to run them in containers. This is a complex process and poses a real challenge. It is like the regular DevOps story of re-architecting a monolith application to be a microservices-based application, only much worse because HPC application libraries tend to be gigabytes and terabytes in size [14]. And that is even before we start discussing all potential security issues (image vulnerabilities, malware, clear text secrets, configuration issues, untrusted images, etc.) [64] or potential performance degradation [65].
  • Even if we manage to package them into containers—which is not a given—we need to be able to run them on a scale, which means running them via Kubernetes. Again, this is not a small task, as understanding Kubernetes architecture, commands, intricacies, and the YAML files we need to create to run applications manually is also very complex.
Regarding workload placement, we must also be extra careful when running HPC applications on a Kubernetes cluster. The way a Kubernetes scheduler distributes workloads leaves much room for improvement. It is well suited for microservice-based web applications, where each component can be run in its container [22]. Still, it is not as efficient as it could be when discussing real-time applications with specific Quality of Service (QoS) requirements [22]. This becomes even more relevant with the size and scale of HPC data centers and their design, as the intelligent scheduling of containers, allocating microservices to containers, container orchestration, elasticity, and automated provisioning play a key role [66]. A vital part of the problem is scalability, primarily due to the software parallelization approach [67]. Containerized HPC applications running in a K8s environment must assume the responsibility for the runtime environment and the system requirement to move between nodes, such as SSH daemons [68]. For example, suppose the only envelope we are pursuing in our HPC data center is the computing power. In that case, we want to have as many many-core processors and accelerators (GPU, FPGA, ASIC) as possible. From the perspective of operating systems used in these environments, there are also some long-forgotten problems with that approach.
For example, let us say that we want to use AMD Epyc Rome or Milan processors in our HPC data center. After we deploy our servers and install operating systems on top of them, we will notice some of the underlying problems we ignored for quite a while re-emerged with a vengeance. As an illustration, if we were to run AMD EPYC 7763 128-thread CPUs on today’s operating systems, we would soon notice that both Linux and Windows are pretty confused about how to use them—they will not recognize their correct core numbers. Sometimes they will require BIOS re-configuration, many BIOS firmware updates and OS updates, etc.
Then, we have other design problems, as our current platforms have a real problem with memory, which is the component that introduces most of the latency into the computing problems (easily two orders of magnitude more significant than what CPU can handle internally). This also leads to the latest CPU designs with integrated memory, which seems to be a move in the right direction.
In that sense, an excellent explanation of why containers are so crucial for HPC is that the demise of Dennard scaling in the early 2000s spurred high-performance computing (HPC) to fully embrace parallelism and rely on Moore’s law for continued performance improvement [69]. This is both good and bad simultaneously—as we cannot stretch out power envelopes to infinity. We mostly thought of this problem as a hardware design problem, but that time is gone now—it is a hardware–software co-design problem. This is especially evident and felt in HPC workloads as these types expose the wrong design patterns even more because of how extreme the HPC workloads are as use cases. This is why we must move forward with our data center design principles to better manage our Kubernetes-based HPC workloads.
That means our hardware-software co-design problem becomes a hardware-software codesign/integration problem. We are no longer in an era where the efficiency of data center design took second place in overall data center performance. That means we must redefine the integration level between our hardware and software to use available resources efficiently, considering hardware heterogeneity and workload diversity in HPC environments.
We have already mentioned some problems related to the Kubernetes scheduler, so let us discuss them. The main issue is the very robotic nature of the default Kubernetes scheduler, which is not suited to latency- and bandwidth-sensitive environments such as HPC. This is why HPC-based environments will need a different architecture for the Kubernetes scheduler to fully use HPC data centers’ capabilities. In that sense, we need more advanced context-aware scheduling algorithms based not only on regular resource utilization metrics such as CPU, RAM, disk, and networks; we need to explore sustainable scheduling techniques in depth to reduce energy consumption and ensure a lower carbon footprint [66]. This is exactly what our software platform does with its hardware-software integration, as it monitors energy usage and feeds that data to multiple AI engines to predict the more optimal placement of workloads when using HPC applications. In Section 6, we will discuss the architecture of our solution to these problems in detail.

5. Experimental Setup and Study Methodology

The experimental setup consisted of four nodes and a Kubernetes control node. Every node has 4 CPU cores and 8 GB of memory and uses SSD for storage. We also reserved a separate node with Intel Core i5-10400F CPU and 16 GB of memory with an NVIDIA GTX 1660 SUPER GPU to train our neural network model. We deliberately chose a simple and light (on hardware) setup as it should indicate whether our methodology works and if our hypothesis stands—the hypothesis being that the Kubernetes scheduler could do better if we fully customized it and if it received help from the ML engine.
Our research started with a general use case—we wanted to see how Kubernetes schedules a regular 3-tier web application via its default scheduler. Then, we embarked on a journey to create our custom Kubernetes scheduler that should do a better job. One of the best examples of how simple the Kubernetes scheduler is that it thinks of a CPU core as a CPU core, no matter the CPU generation, manufacturer, or frequency—it divides every single core into 1000 CPU units.
We performed many rounds of testing for this scenario, starting with ten epochs, going all the way to ten thousand epochs, and performing a statistical analysis of multiple selected epochs. Ultimately, we concluded that five thousand epochs were more than enough to train our ML model as more epochs bring further value, but it is so minuscule that it is not worth our system’s compute time. After running our timing tests, we saved the results to a CSV file. Then, we used that file to train our ML engine to obtain excellent prediction numbers.
The next step was the development of a custom scheduler to iron out the situations where we were not happy with the workload scheduling of the custom Kubernetes scheduler. For that purpose, we used an extended test suite to measure web application response time across multiple scenarios—a scenario in which the default scheduler does the placement and a scenario in which our custom scheduler performs the placement. We wanted to avoid custom schedulers per app at all costs, and ML seemed like an excellent mechanism to do just that—to develop a system based on just one scheduler that takes input from ML for all applications and measures the performance of all of them to have a feedback loop.
Once we performed all these tests for the 3-tier web application scenario with the default and our custom scheduler and confirmed our hypothesis that ML is an excellent replacement for writing custom schedulers per application, we moved to the second stage of our research. That is—to apply the idea of custom scheduling to a test HPC workload. Therefore, we scaled our environment to include more nodes using a set of four HP ProLiant DL380 Gen10 nodes with NVIDIA A100 GPUs. Each node had 24 CPU cores and 384 GB of memory. This time, we used a container image with a set of NVIDIA HPC-Benchmarks to check our performance numbers for a use case much more tailored to HPC. We modified the container to start HPCG-NVIDIA after it booted to obtain reliable data regarding how much the scheduler took before HPCG-NVIDIA began running. We used the same general approach to testing as we did in the web application scenario—of course, the web application was replaced by HPCG-NVIDIA.
For these workload placement decisions to be as correct as possible, we had to create a platform that would use a set of parameters to determine which node was picked based on a score, and this score had to be completely independent of the scores already available in Kubernetes. Also, these parameters needed to be much more nuanced than splitting any CPU core into 1000 CPU units—the parameters needed to reflect the actual performance for a given resource. They needed to represent various hardware devices that contribute to server performance—CPU performance, memory performance, storage performance, networking performance, GPU performance, and FPGA/ASIC availability and, potentially, performance, as it is challenging to develop an actual stack of a priori benchmarks for various FPGA/ASIC accelerators. Power usage (efficiency) needed to be a part of the scoring system, as well as the health state of the server, as we might have wanted to exclude a server from workload placement if there was a hardware failure status attached to it. That meant that all newly added nodes needed to go through multiple phases as they were deployed to the HPC environment—they needed to go through a scanning phase (to determine the hardware content of the server), a performance test phase, and a permanent power and health monitoring phase. These parameters represent a good selection on which to base our scores so we can have enough data to feed to ML engines to learn from. Our research has led us in this direction as it is impossible to achieve proper workload scheduling by counting on the Kubernetes scheduler only, and it is also impossible to create a single scheduler that would work well for all the design ideas we have currently.

6. Testing Results

The first round of our tests was related to a three-tier web application. We measured two sets of data:
  • Default scheduler response time—The time required for the Kubernetes cluster to schedule a three-tier web app and for the front of that app to become available using the default Kubernetes scheduler. This is the correct testing methodology for that scenario, as it is only when the front end starts accepting client requests that we can say that the service is scheduled and operational.
  • Custom scheduler response time—The time required for the Kubernetes cluster to schedule a three-tier web app and for the front end of that app to become available by using our Python-based custom scheduler.
Estimated neural network response time is a neural network prediction for the application response time after training the neural network with a dataset based on custom scheduler workload placement. The overall results of these tests are visible in Table 1.
There is no question that our methodology works and that our hypothesis stands, and the results bear these conclusions out. Our scheduler was always faster than the default one, and the difference reached staggering levels once we reached four nodes, the difference being more than 15%.
The results of these tests gave us an even stronger indicator that our approach works well and merits further research and development. Based on our results, we concluded that there are two ways in which we could proceed with our HPC scenario:
  • We could develop a custom scheduler per HPC application to place applications on Kubernetes pods. This would require a lot of time in terms of coding the Kubernetes schedulers, which might or might not work. Furthermore, there is no guarantee that this approach would work across multiple HPC data centers, making the idea less efficient and potentially wrong.
  • We could develop something that acts as if a set of custom schedulers is present, without writing a set of custom schedulers, that could modify workload placement as it gathers information from the environment, learns about it, and then explicitly places workloads (or better yet, offers to place workloads) on a more suitable node or set of nodes. This is why we went with an ML-based idea that completely replaces the concept of writing multiple schedulers for multiple applications, as this process becomes unnecessary.
After applying the same testing methodology to our HPCG-NVIDIA tests, we obtained the following results (Table 2):
These results solidified our previous conclusion that we could develop an ML-based scheduling scenario that would work well even for resource-hungry applications like HPC. We were also acutely aware that we needed to extend our dataset with data closely related to HPC application performance on a much more general scale. One of the logical premises to that was that overall application performance consists of two phases—fast workload placement and workload placement on the correct node for the application.
The placement of the workload will always be the first step of that process. and the faster that process is, the quicker the calculations will start. But even more importantly, workload placement on the correct node is even more important—after the workload is placed. What good is it to start calculations quickly if containers are not placed on the best available host? We decided to extend our dataset with a large set of additional parameters to depict the HPC use case better, as described in the next section.
We should also mention a couple of related variables that need to be considered when designing a system like this. It is essential to design a Kubernetes environment with an ML engine in mind, meaning at least one server can take the load for the engine. In our first scenario (a three-tier web app with an ML scheduler and training), the cold start time was around 45 s. However, as soon as we started filling in the data for the second scenario with the HPC application in ML, it grew to roughly four minutes as we added a lot of parameters to the decision-making process. This further means that we will have to be very careful in the future in terms of how we train the ML engine (full training from zero or training just with the delta data) and how often we do so. This will further be exacerbated by having hundreds of nodes in our system. Furthermore, this is where the default Kubernetes distributed architecture does not work in our favor—it would be much better to have a completely dedicated management cluster versus the workload cluster like VMware has in VMware Tanzu. It would be better both from the performance and the operational risk perspectives as there are valid reasons why cloud management platforms are usually designed that way, architecture-wise. But in general, there will always be a price to pay here, and we will pay it in the balance of how often we train the model versus the actual benefits of conducting any additional training. We settled on 5000 epochs in our training, which gave us excellent prediction results in a manageable amount of time. It brought only infinitesimally small gains if we increased the number of epochs.

7. Proposed Platform Architecture for Kubernetes Integration with HPC

Dedicated computing systems, like HPC clusters, are preconfigured by systems administrators with specialized software libraries and frameworks. As a result, workloads for these dedicated resources must be built to be hardware-specific and tailored to the system to which it is meant to be deployed to extract the maximum performance [70]. This hinders flexibility if, for example, we want to introduce manageable energy efficiency into the scheduling process. Still, at the same time, we do not want to do so at the expense of environmental complexity and lack of user-friendliness. We also want to have the capability to schedule workloads manually if such a scenario occurs, and we want to achieve all of these scenarios without having to spend a year learning YAML structure and infrastructure as a code, in general, to be able to schedule HPC workloads. For this purpose, it is not sufficient to assign each user a fixed priority (or a static share) and keep it intact [71]. But we can see a tremendous value in using Kubernetes and container platforms to package our applications as—if done correctly—these scenarios will yield better performance than virtual machines and be less wasteful on other resources, like disk space.
Our proposal extends the traditional approach of virtual machines for HPC workloads by providing deeper hardware–software integration. This will involve power usage tracking, deep infrastructure monitoring and scanning, PDU (Power Distribution Unit) management, and using ML engines to learn about the environment and efficiently schedule workloads.
This article proposes a new multi-layer architecture based on the Kubernetes orchestration platform to deploy, configure, manage, and operate a large-scale HPC infrastructure. The only standard architecture and design principle we use is hardware, the same method we always use when designing large-scale HPC centers. We need servers, CPUs, memory, FPGAs, ASICs, and GPUs to handle the workload (what we refer to as the Hardware layer), but how we want that hardware to interact with HPC applications via Kubernetes and how workloads are placed on top of the hardware is entirely different and based on our custom scoring system. Figure 4 illustrates the high-level overview of the architecture of our software platform.
Figure 4. Proposed high-level architecture for using Kubernetes with HPC applications.
Figure 4. Proposed high-level architecture for using Kubernetes with HPC applications.
Electronics 13 02651 g004
Let us now describe this new architecture and the layers it consists of.

7.1. Hardware Stack/Layer

The hardware stack consists of the regular components—PDUs, rack servers, network, InfiniBand and storage switches, storage, and the necessary hardware accelerators extensively used nowadays in modern HPC environments—ASICs, FPGAs, NVIDIA GPUs, etc. No change is needed in the hardware aspect from the regular design patterns, but we will significantly change how we use them via the software layer. 2U and 4U servers offer the best hardware upgradeability via various PCI Express cards, unlike blade servers or most HCI (Hyperconverged Infrastructure) servers, as they are often too limited in expandability and especially hardware compatibility.

7.2. Software Stack/Layer

The critical component of the proposed architecture is a software stack—a combination of a provisioning virtual machine and container-based web application that we use to manage our HPC environment. This software stack is deeply intertwined with hardware in the HPC data center from the ground up, as this is one of the topics where we find that Kubernetes is severely lacking. Kubernetes’s forte is the orchestration of containers without paying much attention to what’s happening below, which is insufficient for what we need in HPC environments. We integrate PDU socket-to-server, storage, network, and network topology mapping into its hardware layer functions. We can see node monitoring in our platform in Figure 5.
Our platform can use simple wizards to integrate existing Docker/Kubernetes environments, whether hosted locally or in the cloud. It can also perform a complete deployment process for supported architectures (x86_64, ARM, and RISC-V) extended with heterogeneous computing components and conduct initial pre-tests to gauge synthetic performance as one of the factors for scheduling decisions. The reason for including this feature is that deploying container technologies like Docker directly for use in an HPC environment is neither trivial nor practical [72], especially on non-x86 platforms. Future iterations of the scheduling process use this information and the information about HPC applications we run in our containers to learn about our environment by inserting ML into scheduling decisions.
The software layer can also scan the underlying hardware to map out its features—CPU models, amount of memory, power supplies and usage, network connections, PCI Express accelerator devices (FPGAs, ASICs, GPUs, controllers), sensors, coolers, etc. It can also monitor the health state of available server components, a capability provided by server remote management options. The software platform then uses these data to learn about the environment via its automated backend and feeds them to ML algorithms. This is necessary as hardware capabilities and health states should be important factors when assessing where to place the HPC workload via operational risk analysis. These data must be fed to the ML engine to refine its scheduling recommendation to the Kubernetes orchestration platform, resulting in better workload placement. This deep hardware and software integration is needed because it yields much better long-term workload placement options that must be easily assigned via a user-friendly UI. As a result, using this deeply integrated architecture, we solve the default Kubernetes scheduler’s lack of finesse and programmability without explicitly developing sets of Kubernetes schedulers for different HPC scenarios and then attach them to Kubernetes pods, which is what we would be forced to do otherwise. There is an excellent example available on the Kubernetes documentation site for multiple scheduler configuration, available at https://kubernetes.io/docs/tasks/extend-kubernetes/configure-multiple-schedulers (accessed on 25 June 2024). It is a fantastic resource to fully grasp the complexity and level of Kubernetes knowledge needed to create it. Of course, we would have to make these custom schedulers per HPC app, which would also translate to statically assigning an HPC app to a specific set of nodes in a cluster if we do not have servers with the same capabilities (same GPU, same FPGA, same ASIC…). On top of the code complexity, this also becomes difficult and complex to manage operationally.
Our platform allows AI engines to suggest where to deploy HPC workloads. Still, we can also override that and explicitly place workloads on any of the Kubernetes nodes managed by our platform. In the workload deployment wizard, we can give our workload a custom name, assign it an image from the registry, and input/output storage location in case we need input and output data storage, as can be seen in Figure 6.
As a part of that process, we also have an approval process when somebody requests manual placement, which can bring about necessary disputes about platform usage. A person in charge of the environment then gets to approve such a request or queue the workload in a fully automated, ML-backed way. Our software layer can also use the built-in QoS mechanisms in Kubernetes to solve some common Kubernetes issues, like poor resource isolation between containers in Kubernetes [73]. One of these mechanisms is the Kubernetes Memory Manager, which is NUMA-aware and will help latency and performance. We can use resource limits on the Kubernetes level (for CPU cores, for example). As mentioned in the previous section, we currently cannot solve direct hardware presentation in containers for all hardware devices, depending on the acceleration technology used (GPU, FPGA, ASIC), but also for more regular hardware, like network cards and storage controllers. Docker, Podman, and Kubernetes are still developing in this direction as this is where the market is pushing them. Until these problems are resolved, there will be workloads that are better suited to virtual machines than containers. That stems from Docker, Podman, and Kubernetes’s original intent—to use CPU and memory as the primary computing fabric, not focusing on additional hardware or accelerators. Consequently, Kubernetes sets two resources for each pod: the central processing unit (CPU) and memory [74].

7.3. Machine Learning Layer

As a new and integral part of our software stack, we integrated multiple ML algorithms to perform real-life research on which algorithm yields the most optimal scheduling and placement results. We deliberately chose to incorporate multiple algorithms to track the progress of these algorithms in terms of their learning and precision on the path of yielding the best scheduling recommendations. The best recommendations in our use case do not necessarily mean only the fastest computing task—this can be brought into context with the amount of power used to perform that computing task (efficiency). From the get-go, our platform to manage HPC workloads on Kubernetes needs to be as efficient and green as possible while meeting the computational requirements in terms of performance, as wasting resources should never be the design goal of any IT system. This is where deep integration between software and hardware comes into play, making scheduling workloads and monitoring environments much more effortless. Furthermore, with health information always available via remote management interfaces, it also makes it easier to calculate scores to determine optimum placement and have alarms available when something on the hardware layer goes wrong. A scoring system we will discuss in this paper is entirely independent of the Kubernetes scheduler’s scoring system, as it lacks finesse for all the different use cases we might encounter in the HPC space. We can see how we implemented ML into our UI in Figure 7.
Furthermore, our software stack needs to be able to override ML suggestions as we cannot start using ML scheduling from the start—we need to give ML time to learn about the environment and applications first. After all, overriding ML suggestions controls our impulses to always put some workloads first at the expense of others. ML algorithms are a valuable tool here as there are QoS parameters that we need to keep in mind when working with various HPC applications—especially on the latency and bandwidth level, depending on the application. We cannot approach this topic from the perspective of just using something simple, such as using response time from the system as the main QoS to be maintained [75].
Our ML layer is packaged as a container and deployed during the infrastructure provisioning phase. We used three distinctive ML types to thoroughly compare workload scheduling and placement recommendations without bias. That way, we had multiple datasets to compare placement decisions as we developed our platform. Specifically, we used DNN (Deep Neural Network), DF (Decision Forest), and LR (linear regression). We integrated them into the UI of our platform so that we could quickly get to the corresponding scores.

7.4. User Interface Layer

The UI-based part of the application needs to be usable by a regular user, not an IT expert. Therefore, it needs to combine the simplicity of drag-and-drop wizards with the complexity of Docker and Podman in the backend. We also created a prompt-based interface (a custom Discord chatbot application) to interact with the backend ML engine to check the state of our environment or schedule workloads, even if a scientist who wants to start an HPC workload is not sitting at the computer. This makes it very straightforward to schedule/manage workloads without spending hours defining how and where these tasks should be started. In Figure 8, we can see information about the current state of our environment provided by the chatbot:
One of the biggest challenges in making such a user-friendly interface is that we need to quickly package various HPC applications in containers without writing many YAML configuration files. Our platform implemented these features via multi-arch base containers and multi-arch applications, presented as Docker image layers, that can be imported as files or by working with our Harbor registry. Harbor is an open-source registry that can be used to store, distribute, sign, and scan container images. Multi-arch base containers (Ubuntu, CentOS, or Debian for x86, ARM, or RISC-V) are a good choice as they are small enough not to waste resources while still being supported by most HPC applications. In HPC environments based on containers, it is essential to have the capability to run the HPC application instead of insisting on using the most common small container Linux distributions (for example, Alpine Linux), as these require additional Linux packages [18]. That also means we had to pre-build our base containers with specific requirements tailored to HPC environments. For example, they must support high-performance libraries like MPI and CUDA for NVIDIA GPU-based applications, etc. The HPC application container will include the application and its dependencies, such as the MPI library, accelerator libraries and tools, and parallel math libraries [19]. Then, we have a drag-and-drop system to automate container building with additional layers for HPC applications. This way, we avoid writing YAML files and forcing users to build container images manually. We should use the same design principle for storage and network integration—we integrate storage and network at the platform level and use storage/networking as input and output variables in the wizard so that they can be easily re-used by Kubernetes API when making API calls. This also enables us to integrate principles of micro-segmentation to reduce attack surfaces on our HPC applications and secure them as much as possible.
We can see the wizard-driven approach that we took for adding storage in Figure 9, as well as the storage types supported:
This storage wizard not only simplifies the process but also saves valuable time for our users, allowing them to focus on their core tasks, such as placing workloads on Kubernetes clusters. There’s no need to juggle multiple commands or learn additional parameters, as the wizard handles everything efficiently.

7.5. Monitoring Layer

Apart from hardware monitoring for the scoring system, we also need to efficiently monitor what is happening in our environment from the management perspective, i.e., via dashboards and visual data representation. This is where integration with the backend Clickhouse database makes much sense, as it is an extensive, column-sorted analytics database. The increasing complexity of future computing systems requires monitoring and management tools to provide a unified view of the center, perfect for understanding the hardware and general performance aspects of an extreme cluster or high-performance system [20]. For example, we can see how we integrated PDU monitoring into our platform in Figure 10.
This also allows us to prototype and embed additional graphs quickly, should the need arise. We can easily collect a large amount of data that needs to be processed [76].

8. Scheduling of HPC Workloads on Our Platform via ML or Manual Placement

In the previous section, we mentioned problems with the Kubernetes scheduler, the complexity of creating a custom one, and how our platform avoids that process altogether. We achieved this by allowing the AI engine to explicitly start an HPC application packaged as a container on the node that it deems the most suitable for it if we choose to do so. The complex process creates a container image and its corresponding layers to package the application. We wanted to do away with the complexity of creating images, as this will significantly impact how we resolve software dependencies in HPC applications, as resolving lifecycle dependencies at the container orchestration level adds additional overhead [77]. The way we solve these problems is by using a wizard-driven process based on a drag-and-drop canvas where we can drop storage locations, base containers, HPC applications, and networks—generally speaking, all of the necessary details to create a workload with all of the settings required to run on a Kubernetes cluster. That means that, during this process, we go through four phases, as shown in Figure 11.
These four phases in our wizard ensure that we have all the necessary details to start the workload and all the prerequisites for it to run successfully.
There are subtle differences in manual workload placement compared to ML-based placement, so let us delve deeper into that.

8.1. Manual Workload Placement

Manual workload placement is simple—from the list of available nodes, we ignore everything that ML provides to us and click on the “Deploy” button on the right side of the wizard for the node that we want to select for workload deployment, as we can see in Figure 12.
This means that if we—for whatever reason—want to manually select a node chosen by the ML and presented via scores (to be discussed in the following sub-section), we can easily do so. We can also see that scores are visible in the UI no matter which deployment model we choose. Manual workload placement might be beneficial if we have a set of HPC nodes reserved for some emergency workload or if we are starting to use some new HPC application and want to gauge its performance.

8.2. ML-Based Workload Placement

ML-based workload placement in our platform is based on a set of parameters that we determined to be crucial for placement decisions to be as accurate and reliable as possible. Whichever workload placement methodology we choose, our application consults with the ML algorithms in the backend before making scheduling decisions. The ML engine considers the parameter history on a per-app, per-server basis. There are five sets of parameters:
  • POWER usage—A set of timestamped power readings from the PDU socket and server remote management, sampled at configurable intervals. These readings estimate HPC application power requirements, significantly improving the platform’s energy efficiency, especially given time and many workload executions to become even more accurate.
  • HEALTH information—A set of parameters taken from server remote management handling the health states of components, specifically fans, memory health state, power supply health state, power state (redundant or not), processor health state, storage health state, network, and remote management health state and temperatures.
  • CPU, memory, storage, and networking testing results—As the server is provisioned from our platform in multiple passes, pre-determined synthetic and real-life benchmarks are automatically performed and averaged across the configurable number of runs. These bare-metal and containerized benchmarks determine the baseline hardware performance level for all servers added to the system. We use sysbench, stress-ng, hdparm, HPL, HPCC, and HPL with various parameters to gauge performance in single and multi-tasking scenarios.
  • NVIDIA GPU results—A set of pre-determined synthetic and real-life benchmarks is automatically performed. The server with the installed NVIDIA GPU is provisioned from our platform in multiple takes, averaged across the configurable number of runs. We use NVIDIA HPC-Benchmarks for this purpose.
  • FPGA/ASIC availability—For supported FPGA/ASIC controllers, platform users can manually add additional scores per app. Ideally, this would be automated, but it is currently impossible because of the different software stacks used by FPGAs and ASICs.
All scores are taken per application, not as an aggregate, and our scoring system takes current usage as a parameter into consideration, as well. For example, to avoid operational risk, if two servers have the same NVIDIA GPU, one of those GPUs is being used less at a certain point in time, and our ML scheduler automatically picks the server with less GPU used by other apps. This enables us to have individualized, per-app scores that are much more relevant as we develop the platform for general use cases that support even more hardware. The numbered scoring system considers all of these parameters but also dispatches negative grades if there are different types of failures. For example, if one of the power supplies on the server fails, that server receives a score downgrade. Simplified, the algorithm works as shown in Figure 13:
All the mentioned parameters have weights that can be used as preset values (provided by the app itself, as fixed values) or changed by the platform manager by editing values assigned to any of these parameters. That, in turn, changes the server’s overall per-application score and the relative score related to the set of parameters we changed manually. This seems like a reasonable compromise between automation and the capability to override the system, as that option always needs to exist.
Hardware tests are executed post-deployment before any workloads can be placed via our platform. Ansible playbooks in our deployment process automatically install and perform all necessary tests (multiple runs). Scores are saved per node in a CSV file, which is then added to the Clickhouse database and uploaded to ML. ML engines always have up-to-date data when we expand our environment.
All workload placement processes via our platform are also “captured” by ML using the same method as the initial one, enabling us to track performance and load on our servers constantly. By correlating power usage and health info data, ML engines also have up-to-date scores for every server added to our platform. Scores are automatically calculated and added next to each node’s name, making it easier for users to track the most recent scores. Suppose we do not manually select any specific node. In that case, these scores are used as an a priori recommendation system that automatically places the workload on a node with the highest score.

9. Future Work

This paper presents our first steps towards exploiting virtualization and containerization technology in specifics of heterogeneous HPC environments. Multiple routes exist for developing our proposed platform to handle HPC workloads via Kubernetes; the first is to integrate capabilities to run virtualized workloads for some use cases.
We mentioned that there are and likely will be workloads for which virtual machines are much better suited than containers, at least for now. That means that some integration with a purely hypervisor-based virtualized platform is a good development path for the future. Specifically, it needs to be as automated and easy to use as the platform we described and integrate various technologies for various operating systems and solutions. For example, cloud-init and cloudbase-init could customize virtual machines post-deployment from a template. After that, PowerShell and Ansible could be used to deploy and configure the necessary software components. With the current state of technology, it would not be all that difficult to integrate that type of platform with some more advanced settings that are currently not viable solutions with Kubernetes, like direct hardware presentation via PCIe passthrough or something like SR-IOV, to partition PCIe cards into multiple virtual functions and provide them automatically to virtual machines. That would make it much easier to use FPGAs and ASICs that need to be fixed with Kubernetes (unfortunately, that list is quite long). We could then look into adding NUMA awareness to VM-based workloads, as this is where it would be most beneficial.
One prominent research area for the future is creating a stack of automated procedures for a priori performance evaluation of FPGAs and ASICs for baseline purposes. The current limitation of our approach is that it is limited to NVIDIA GPUs only, as we were able to develop a baseline methodology that works and is repeatable. FPGA/ASIC performance evaluation is currently not feasible as complexities and software availability bring various licensing and copyright issues, as quite a few of the required software components are hidden behind a paywall or have complex download and deployment procedures. However, if we look at the automation process, this is not an impossible challenge.
Another limitation of our research is that it does not evaluate storage or networking performance, at least for now. We are currently developing this, which will be integrated into the final version of our software stack. When we finish, it will be easy to integrate with the current evaluation as our evaluation algorithm is modular, and we can easily add parameters. We know that networking performance (both latency and bandwidth) plays a vital role in HPC application performance, as does storage, depending on the use case. This integration will add another level of data to our training baseline, further improving our ML scheduler’s decision-making process. But as it stands, our ML scheduler is much more precise than anything Kubernetes has natively.
The next area of future research might be creating an orchestrated methodology to automate advanced capabilities like SR-IOV and PCIe forwarding but from the physical perspective. We could then integrate that into our design, giving us some advantages regarding available bandwidth and better latency. As mentioned in this paper, many hardware devices are incompatible with Kubernetes due to a lack of drivers that could assign them directly to Kubernetes pods. We could partially work around this problem by orchestrating configuration from the physical perspective and assigning them to container resources. An example of this is advanced networking for HPC workloads. For example, let us say we want to use two link-aggregated 25 Gbit/s network cards from the perspective of the Kubernetes pod. We would typically do so to increase the bandwidth available to a Kubernetes pod. These configurations are difficult to achieve on the Kubernetes level, but they are achievable with Ansible on the physical level. Then, the Kubernetes layer could sit on top of this configuration. Previous research shows many opportunities to optimize Kubernetes further networking when working with HPC workloads, as Kubernetes presents overheads for several HPC applications over TCP/IP protocol [78].
We could then research developing an ML-based scheduler that is application-aware to spool up containers or virtual machines, depending on the available hardware, accelerators, applications, and their requirements. Furthermore, even more developed LLM-based (Large Language Model) management via a Discord bot could become even more helpful for migrating workloads or application states from one node to another.

10. Conclusions

This paper proposes a novel platform to manage Kubernetes-based environments for HPC. It also covers a broad range of topics related to containerization, virtualization, orchestration, and HPC topics. When discussing the specifics of HPC, we concluded that using Kubernetes and containers for HPC workloads is complex and needs to be simplified as much as possible. This is why we developed a new platform with proactive and reactive components. It can proactively assign specific performance numbers to our HPC data center servers and reactively help place HPC workloads on them. We concluded that there are better ways to go than developing custom Kubernetes schedulers on a per-HPC application basis. Instead, we created a platform that uses ML to load workloads automatically and manually. This approach took a lot of integration of various components—PDUs to track energy usage, remote management interfaces to track servers’ health state, etc. All these data are fed to the ML engine to predict and suggest which Kubernetes node in our cluster is the optimal node to put that workload. Our study shows much room for improvement in how we place our workloads in Kubernetes (in general), especially in HPC environments. Our platform offers a simple opportunity to go down that path without much additional effort, which we set out to do when we started this research.

Author Contributions

Conceptualization, V.D.; Methodology, V.D.; Validation, V.D. and J.S.; Formal analysis, V.D. and J.S.; Investigation, J.S.; Resources, V.D.; Writing—original draft, V.D.; Writing—review & editing, V.D.; Supervision, M.K.; Project administration, M.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Đorđević, B.; Kraljević, N.; Davidović, N. Performance Comparison of CPU Hardware-Assisted Features for the Type-2 Hypervisors. In Proceedings of the 2024 23rd International Symposium INFOTEH-JAHORINA (INFOTEH), Jahorina, Bosnia and Herzegovina, 20–22 March 2024. [Google Scholar] [CrossRef]
  2. Chen, Y.-R.; Liu, I.-H.; Chou, C.-W.; Li, J.-S.; Liu, C.-G. Multiple Virtual Machines Live Migration Scheduling Method Study on VMware vMotion. In Proceedings of the 2018 3rd International Conference on Computer and Communication Systems (ICCCS), Nagoya, Japan, 27–30 April 2018. [Google Scholar] [CrossRef]
  3. Shirinbab, S.; Lundberg, L.; Hakansson, J. Comparing Automatic Load Balancing Using VMware DRS with a Human Expert. In Proceedings of the 2016 IEEE International Conference on Cloud Engineering Workshop (IC2EW), Berlin, Germany, 4–8 April 2016. [Google Scholar] [CrossRef]
  4. Li, Z.; Kihl, M.; Lu, Q.; Andersson, J.A. Performance Overhead Comparison between Hypervisor and Container Based Virtualization. In Proceedings of the 2017 IEEE 31st International Conference on Advanced Information Networking and Applications (AINA), Taipei, Taiwan, 27–29 March 2017. [Google Scholar] [CrossRef]
  5. Wang, P.; Posey, S. GPU Best Practices for HPC Applications at Industry Scale. In GPU Solutions to Multi-Scale Problems in Science and Engineering; Lecture Notes in Earth System Sciences; Springer: Berlin/Heidelberg, Germany, 2013; pp. 163–172. [Google Scholar] [CrossRef]
  6. Nonaka, J.; Ono, K.; Fujita, M. 234Compositor: A Flexible Parallel Image Compositing Framework for Massively Parallel Visualization Environments. Future Gener. Comput. Syst. 2018, 82, 647–655. [Google Scholar] [CrossRef]
  7. Vu, D.-D.; Tran, M.-N.; Kim, Y. Predictive Hybrid Autoscaling for Containerized Applications. IEEE Access 2022, 10, 109768–109778. [Google Scholar] [CrossRef]
  8. Milroy, D.J.; Misale, C.; Georgakoudis, G.; Elengikal, T.; Sarkar, A.; Drocco, M.; Patki, T.; Yeom, J.-S.; Gutierrez, C.E.A.; Ahn, D.H.; et al. One Step Closer to Converged Computing: Achieving Scalability with Cloud-Native HPC. In Proceedings of the 2022 IEEE/ACM 4th International Workshop on Containers and New Orchestration Paradigms for Isolated Environments in HPC (CANOPIE-HPC), Dallas, TX, USA, 14 November 2022; pp. 57–70. [Google Scholar] [CrossRef]
  9. Lublinsky, B.; Jennings, E.; Spišaková, V. A Kubernetes ‘Bridge’ Operator between Cloud and External Resources. In Proceedings of the 2023 8th International Conference on Cloud Computing and Big Data Analytics (ICCCBDA), Chengdu, China, 26–28 April 2023; pp. 263–269. [Google Scholar] [CrossRef]
  10. Spišaková, V.; Klusáček, D.; Hejtmánek, L. Using Kubernetes in Academic Environment: Problems and Approaches (Open Scheduling Problem). Available online: https://jsspp.org/papers22/6.pdf (accessed on 22 May 2024).
  11. Lingayat, A.; Badre, R.R.; Kumar Gupta, A. Performance Evaluation for Deploying Docker Containers on Baremetal and Virtual Machine. In Proceedings of the 2018 3rd International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 15–16 October 2018. [Google Scholar] [CrossRef]
  12. Agarwal, K.; Jain, B.; Porter, D.E. Containing the Hype. In Proceedings of the 6th Asia-Pacific Workshop on Systems, Tokyo, Japan, 27–28 July 2015. [Google Scholar] [CrossRef]
  13. Antunes, C.; Vardasca, R. Performance of Jails versus Virtualization for Cloud Computing Solutions. Procedia Technol. 2014, 16, 649–658. [Google Scholar] [CrossRef]
  14. Trigo, A.; Varajão, J.; Sousa, L. DevOps Adoption: Insights from a Large European Telco. Cogent Eng. 2022, 9, 2083474. [Google Scholar] [CrossRef]
  15. Soltesz, S.; Pötzl, H.; Fiuczynski, M.E.; Bavier, A.; Peterson, L. Container-based operating system virtualization: A scalable, high-performance alternative to hypervisors. In Proceedings of the 2nd ACM SIGOPS/EuroSys European Conference on Computer Systems 2007, Lisbon, Portugal, 21–23 March 2007; Volume 41, pp. 275–287. [Google Scholar] [CrossRef]
  16. Li, X.; Jiang, J. Performance Analysis of PaaS Cloud Resources Management Model Based on LXC. In Proceedings of the 2016 International Conference on Cloud Computing and Internet of Things (CCIOT), Dalian, China, 22–23 October 2016; pp. 118–130. [Google Scholar] [CrossRef]
  17. Younge, A.J.; Pedretti, K.; Grant, R.E.; Brightwell, R. A Tale of Two Systems: Using Containers to Deploy HPC Applications on Supercomputers and Clouds. In Proceedings of the 2017 IEEE International Conference on Cloud Computing Technology and Science (CloudCom), Hong Kong, China, 11–14 December 2017; pp. 74–81. [Google Scholar] [CrossRef]
  18. Zhang, X.; Li, L.; Wang, Y.; Chen, E.; Shou, L. Zeus: Improving Resource Efficiency via Workload Colocation for Massive Kubernetes Clusters. IEEE Access 2021, 9, 105192–105204. [Google Scholar] [CrossRef]
  19. Felter, W.; Ferreira, A.; Rajamony, R.; Rubio, J. An Updated Performance Comparison of Virtual Machines and Linux Containers. In Proceedings of the 2015 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS), Philadelphia, PA, USA, 29–31 March 2015; pp. 171–172. [Google Scholar] [CrossRef]
  20. Burns, B.; Grant, B.; Oppenheimer, D.; Brewer, E.; Wilkes, J. Borg, Omega, and Kubernetes. Queue 2016, 14, 70–93. [Google Scholar] [CrossRef]
  21. Dordevic, B.; Timcenko, V.; Lazic, M.; Davidovic, N. Performance Comparison of Docker and Podman Container-Based Virtualization. In Proceedings of the 2022 21st International Symposium INFOTEH-JAHORINA (INFOTEH), East Sarajevo, Bosnia and Herzegovina, 16–18 March 2022. [Google Scholar] [CrossRef]
  22. Gantikow, H.; Walter, S.; Reich, C. Rootless Containers with Podman for HPC. In High Performance Computing, Proceedings of the International Conference on High Performance Computing, Frankfurt am Main, Germany, 22–25 June 2020; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 21–25 June 2020; pp. 343–354. [Google Scholar] [CrossRef]
  23. Sheka, A.; Bersenev, A.; Samun, V. Containerization in Scientific Calculations. In Proceedings of the 2019 International Multi-Conference on Engineering, Computer and Information Sciences (SIBIRCON), Novosibirsk, Russia, 21–27 October 2019. [Google Scholar] [CrossRef]
  24. Kiourtis, A.; Karabetian, A.; Karamolegkos, P.; Poulakis, Y.; Mavrogiorgou, A.; Kyriazis, D. A Comparison of Container Systems for Machine Learning Scenarios: Docker and Podman. In Proceedings of the 2022 2nd International Conference on Computers and Automation (CompAuto), Paris, France, 18–20 August 2022. [Google Scholar] [CrossRef]
  25. Stephey, L.; Canon, S.; Gaur, A.; Fulton, D.; Younge, A.J. Scaling Podman on Perlmutter: Embracing a Community-Supported Container Ecosystem. In Proceedings of the 2022 IEEE/ACM 4th International Workshop on Containers and New Orchestration Paradigms for Isolated Environments in HPC (CANOPIE-HPC), Dallas, TX, USA, 4 November 2022. [Google Scholar] [CrossRef]
  26. Khaleq, A.A.; Ra, I. Intelligent Autoscaling of Microservices in the Cloud for Real-Time Applications. IEEE Access 2021, 9, 35464–35476. [Google Scholar] [CrossRef]
  27. Bernstein, D. Containers and Cloud: From LXC to Docker to Kubernetes. IEEE Cloud Comput. 2014, 1, 81–84. [Google Scholar] [CrossRef]
  28. Jagadeeswari, N.; Mohanraj, V.; Suresh, Y.; Senthilkumar, J. Optimization of Virtual Machines Performance Using Fuzzy Hashing and Genetic Algorithm-Based Memory Deduplication of Static Pages. Automatika 2023, 64, 868–877. [Google Scholar] [CrossRef]
  29. Lee, G. High-Performance Computing Networks. In Cloud Networking; Elsevier B.V.: Amsterdam, The Netherlands, 2014; pp. 179–189. [Google Scholar] [CrossRef]
  30. Yang, H.; Ong, S.K.; Nee, A.Y.C.; Jiang, G.; Mei, X. Microservices-Based Cloud-Edge Collaborative Condition Monitoring Platform for Smart Manufacturing Systems. Int. J. Prod. Res. 2022, 60, 7492–7501. [Google Scholar] [CrossRef]
  31. Holmes, V.; Newall, M. HPC and the Big Data Challenge. Saf. Reliabil. 2016, 36, 213–224. [Google Scholar] [CrossRef]
  32. Houzeaux, G.; Garcia-Gasulla, M. High Performance Computing Techniques in CFD. Int. J. Comput. Fluid Dyn. 2020, 34, 457. [Google Scholar] [CrossRef]
  33. Örmecioğlu, T.O.; Aydoğdu, İ.; Örmecioğlu, H.T. GPU-Based Parallel Programming for FEM Analysis in the Optimization of Steel Frames. J. Asian Archit. Build. Eng. 2024, 2024, 2345310. [Google Scholar] [CrossRef]
  34. Jha, A.V.; Teri, R.; Verma, S.; Tarafder, S.; Bhowmik, W.; Kumar Mishra, S.; Appasani, B.; Srinivasulu, A.; Philibert, N. From Theory to Practice: Understanding DevOps Culture and Mindset. Cogent Eng. 2023, 10, 2251758. [Google Scholar] [CrossRef]
  35. Li, H.; Kettinger, W.J.; Yoo, S. Dark Clouds on the Horizon? Effects of Cloud Storage on Security Breaches. J. Manag. Inf. Syst. 2024, 41, 206–235. [Google Scholar] [CrossRef]
  36. Greneche, N.; Cerin, C. Autoscaling of Containerized HPC Clusters in the Cloud. In Proceedings of the 2022 IEEE/ACM International Workshop on Interoperability of Supercomputing and Cloud Technologies (SuperCompCloud), Dallas, TX, USA, 13–18 November 2022; pp. 1–7. [Google Scholar] [CrossRef]
  37. Liu, P.; Guitart, J. Fine-Grained Scheduling for Containerized HPC Workloads in Kubernetes Clusters. Available online: http://arxiv.org/abs/2211.11487 (accessed on 22 May 2024).
  38. Beltre, A.M.; Saha, P.; Govindaraju, M.; Younge, A.; Grant, R.E. Enabling HPC Workloads on Cloud Infrastructure Using Kubernetes Container Orchestration Mechanisms. In Proceedings of the 2019 IEEE/ACM International Workshop on Containers and New Orchestration Paradigms for Isolated Environments in HPC (CANOPIE-HPC), Denver, CO, USA, 18 November 2019; pp. 11–20. [Google Scholar] [CrossRef]
  39. Hursey, J. A Separated Model for Running Rootless, Unprivileged PMIx-Enabled HPC Applications in Kubernetes. In Proceedings of the 2022 IEEE/ACM 4th International Workshop on Containers and New Orchestration Paradigms for Isolated Environments in HPC (CANOPIE-HPC), Dallas, TX, USA, 14 November 2022. [Google Scholar]
  40. Jang, H.-C.; Luo, S.-Y. Enhancing Node Fault Tolerance through High-Availability Clusters in Kubernetes. In In Proceedings of the 2023 IEEE 3rd International Conference on Electronic Communications, Internet of Things and Big Data (ICEIB) 2023, Taichung, Taiwan, 14–15 April 2023. [Google Scholar] [CrossRef]
  41. Ding, Z.; Wang, S.; Jiang, C. Kubernetes-Oriented Microservice Placement with Dynamic Resource Allocation. IEEE Trans. Cloud Comput. 2023, 11, 1777–1793. [Google Scholar] [CrossRef]
  42. Hursey, J. Design Considerations for Building and Running Containerized MPI Applications. In Proceedings of the 2020 2nd International Workshop on Containers and New Orchestration Paradigms for Isolated Environments in HPC (CANOPIE-HPC), Atlanta, GA, USA, 12 November 2020. [Google Scholar] [CrossRef]
  43. Sukhija, N.; Bautista, E. Towards a Framework for Monitoring and Analyzing High Performance Computing Environments Using Kubernetes and Prometheus. In Proceedings of the 2019 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computing, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI), Leicester, UK, 19–23 August 2019; pp. 257–262. [Google Scholar] [CrossRef]
  44. Kosinska, J.; Tobiasz, M. Detection of Cluster Anomalies with ML Techniques. IEEE Access 2022, 10, 110742–110753. [Google Scholar] [CrossRef]
  45. Sebrechts, M.; Borny, S.; Wauters, T.; Volckaert, B.; De Turck, F. Service Relationship Orchestration: Lessons Learned from Running Large Scale Smart City Platforms on Kubernetes. IEEE Access 2021, 9, 133387–133401. [Google Scholar] [CrossRef]
  46. Vasireddy, I.; Ramya, G.; Kandi, P. Kubernetes and Docker Load Balancing: State-of-the-Art Techniques and Challenges. Int. J. Innov. Res. Eng. Manag. 2023, 10, 49–54. [Google Scholar] [CrossRef]
  47. Vohra, D. Installing Kubernetes Using Docker. In Kubernetes Microservices with Docker; Springer: Berlin/Heidelberg, Germany, 2016; pp. 3–38. [Google Scholar] [CrossRef]
  48. Liu, B.; Li, J.; Lin, W.; Bai, W.; Li, P.; Gao, Q. K-PSO: An Improved PSO-based Container Scheduling Algorithm for Big Data Applications. Int. J. Netw. Manag. 2020, 31, e2092. [Google Scholar] [CrossRef]
  49. Malviya, A.; Dwivedi, R.K. A Comparative Analysis of Container Orchestration Tools in Cloud Computing. In Proceedings of the 2022 9th International Conference on Computing for Sustainable Global Development (INDIACom) 2022, New Delhi, India, 24–25 March 2022. [Google Scholar] [CrossRef]
  50. Pan, Y.; Chen, I.; Brasileiro, F.; Jayaputera, G.; Sinnott, R. A Performance Comparison of Cloud-Based Container Orchestration Tools. In Proceedings of the 2019 IEEE International Conference on Big Knowledge (ICBK), Beijing, China, 10–11 November 2019. [Google Scholar] [CrossRef]
  51. Lee, S.; Raza Shah, S.A.; Seok, W.; Moon, J.; Kim, K.; Raza Shah, S.H. An Optimal Network-Aware Scheduling Technique for Distributed Deep Learning in Distributed HPC Platforms. Electronics 2023, 12, 3021. [Google Scholar] [CrossRef]
  52. Zha, B.; Shen, H. Adaptively Periodic I/O Scheduling for Concurrent HPC Applications. Electronics 2022, 11, 1318. [Google Scholar] [CrossRef]
  53. Granhão, D.; Canas Ferreira, J. Transparent Control Flow Transfer between CPU and Accelerators for HPC. Electronics 2021, 10, 406. [Google Scholar] [CrossRef]
  54. Ruhela, A.; Xu, S.; Manian, K.V.; Subramoni, H.; Panda, D.K. Analyzing and Understanding the Impact of Interconnect Performance on HPC, Big Data, and Deep Learning Applications: A Case Study with InfiniBand EDR and HDR. In Proceedings of the 2020 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) 2020, New Orleans, LA, USA, 18–22 May 2020. [Google Scholar] [CrossRef]
  55. Aly, M.; Khomh, F.; Yacout, S. Kubernetes or OpenShift? Which Technology Best Suits Eclipse Hono IoT Deployments. In Proceedings of the 2018 IEEE 11th Conference on Service-Oriented Computing and Applications (SOCA), Paris, France, 20–22 November 2018. [Google Scholar] [CrossRef]
  56. Linzel, B.; Zhu, E.; Flores, G.; Liu, J.; Dikaleh, S. How can OpenShift accelerate your Kubernetes adoption: A workshop exploring OpenShift features. In Proceedings of the CASCON’19: Proceedings of the 29th Annual International Conference on Computer Science and Software Engineering, Markham, ON, Canada, 4–6 November 2019; pp. 380–381. [Google Scholar] [CrossRef]
  57. Vohra, D. Using an HA Master with OpenShift. In Kubernetes Management Design Patterns; Apress: New York, NY, USA, 2017; pp. 335–353. [Google Scholar] [CrossRef]
  58. Marksteiner, P. High-Performance Computing—An Overview. Comput. Physics Commun. 1996, 97, 16–35. [Google Scholar] [CrossRef]
  59. Cardoso, J.M.P.; Coutinho, J.G.F.; Diniz, P.C. High-Performance Embedded Computing. In Proceedings of the Embedded Computing for High Performance, Waltham, MA, USA, 12–14 September 2017; pp. 17–56. [Google Scholar] [CrossRef]
  60. Feng, W.; Manocha, D. High-Performance Computing Using Accelerators. Parallel Comput. 2007, 33, 645–647. [Google Scholar] [CrossRef]
  61. Kindratenko, V.; Thiruvathukal, G.K.; Gottlieb, S. High-Performance Computing Applications on Novel Architectures. Comput. Sci. Eng. 2008, 10, 13–15. [Google Scholar] [CrossRef]
  62. Lee, V.W.; Grochowski, E.; Geva, R. Performance Benefits of Heterogeneous Computing in HPC Workloads. In Proceedings of the 2012 IEEE 26th International Parallel and Distributed Processing Symposium Workshops & PhD Forum 2012, Shanghai, China, 21–25 May 2012. [Google Scholar] [CrossRef]
  63. Ambrosino, G.; Fioccola, G.B.; Canonico, R.; Ventre, G. Container Mapping and Its Impact on Performance in Containerized Cloud Environments. In Proceedings of the 2020 IEEE International Conference on Service Oriented Systems Engineering (SOSE), Oxford, UK, 3–6 August 2020. [Google Scholar] [CrossRef]
  64. Senjab, K.; Abbas, S.; Ahmed, N.; Khan, A.U.R. A Survey of Kubernetes Scheduling Algorithms. J. Cloud Comput. 2023, 12, 87. [Google Scholar] [CrossRef]
  65. Carrión, C. Kubernetes Scheduling: Taxonomy, Ongoing Issues and Challenges. ACM Comput. Surv. 2022, 55, 1–37. [Google Scholar] [CrossRef]
  66. Rodriguez, G.; Yannibelli, V.; Rocha, F.G.; Barbara, D.; Azevedo, I.M.; Menezes, P.M. Understanding and Addressing the Allocation of Microservices into Containers: A Review. IETE J. Res. 2023, 1–14. [Google Scholar] [CrossRef]
  67. Gomez, C.; Martinez, F.; Armejach, A.; Moreto, M.; Mantovani, F.; Casas, M. Design Space Exploration of Next-Generation HPC Machines. In Proceedings of the 2019 IEEE International Parallel and Distributed Processing Symposium (IPDPS), Rio de Janeiro, Brazil, 20–24 May 2019. [Google Scholar] [CrossRef]
  68. Tesliuk, A.; Bobkov, S.; Ilyin, V.; Novikov, A.; Poyda, A.; Velikhov, V. Kubernetes Container Orchestration as a Framework for Flexible and Effective Scientific Data Analysis. In Proceedings of the 2019 Ivannikov Ispras Open Conference (ISPRAS), Moscow, Russia, 5–6 December 2019; pp. 67–71. [Google Scholar] [CrossRef]
  69. Rathmayer, S.; Lenke, M. A Tool for On-Line Visualization and Interactive Steering of Parallel HPC Applications. In Proceedings of the 11th International Parallel Processing Symposium, Geneva, Switzerland, 1–5 April 1997. [Google Scholar] [CrossRef]
  70. Ruiz, L.M.; Pueyo, P.P.; Mateo-Fornes, J.; Mayoral, J.V.; Tehas, F.S. Autoscaling Pods on an On-Premise Kubernetes Infrastructure QoS-Aware. IEEE Access 2022, 10, 33083–33094. [Google Scholar] [CrossRef]
  71. Lossent, A.; Rodriguez Peon, A.; Wagner, A. PaaS for Web Applications with OpenShift Origin. J. Phys. Conf. Ser. 2017, 898, 082037. [Google Scholar] [CrossRef]
  72. Levesque, J.; Wagenbreth, G. High Performance Computing; Chapman and Hall/CRC: Boca Raton, FL, USA, 2010. [Google Scholar] [CrossRef]
  73. Souppaya, M.; Morello, J.; Scarfone, K. Application Container Security Guide; National Institute of Standards and Technology: Gaithersburg, MD, USA, 2017. [Google Scholar] [CrossRef]
  74. Flora, J.; Goncalves, P.; Teixeira, M.; Antunes, N. A Study on the Aging and Fault Tolerance of Microservices in Kubernetes. IEEE Access 2022, 10, 132786–132799. [Google Scholar] [CrossRef]
  75. Zhou, N.; Georgiou, Y.; Zhong, L.; Zhou, H.; Pospieszny, M. Container Orchestration on HPC Systems. In Proceedings of the 2020 IEEE 13th International Conference on Cloud Computing (CLOUD), Virtual Event, 18–24 October 2020; pp. 34–36. [Google Scholar] [CrossRef]
  76. Roslin Dayana, K.; Shobha Rani, P. Secure Cloud Data Storage Solution with Better Data Accessibility and Time Efficiency. Automatika 2023, 64, 756–763. [Google Scholar] [CrossRef]
  77. Grigoryan, G.; Kwon, M.; Rafique, M.M. Extending the Control Plane of Container Orchestrators for I/O Virtualization. In Proceedings of the 2020 2nd International Workshop on Containers and New Orchestration Paradigms for Isolated Environments in HPC (CANOPIE-HPC), Atlanta, GA, USA, 12 November 2020. [Google Scholar] [CrossRef]
  78. Smith, M.C.; Drager, S.L.; Pochet, L.; Peterson, G.D. High Performance Reconfigurable Computing Systems. In Proceedings of the 44th IEEE 2001 Midwest Symposium on Circuits and Systems. MWSCAS 2001 (Cat. No.01CH37257), Dayton, OH, USA, 14–17 August 2001. [Google Scholar] [CrossRef]
Figure 1. The architecture of Docker and its most essential services, https://commons.wikimedia.org/wiki/File:Docker-architecture.png, accessed on 22 June 2024.
Figure 1. The architecture of Docker and its most essential services, https://commons.wikimedia.org/wiki/File:Docker-architecture.png, accessed on 22 June 2024.
Electronics 13 02651 g001
Figure 2. Podman has a much simplified and less monolithic service architecture than Docker.
Figure 2. Podman has a much simplified and less monolithic service architecture than Docker.
Electronics 13 02651 g002
Figure 3. Kubernetes service architecture includes control plane services and services on hosts running the workloads Kubernetes manages.
Figure 3. Kubernetes service architecture includes control plane services and services on hosts running the workloads Kubernetes manages.
Electronics 13 02651 g003
Figure 5. Node monitoring in our platform.
Figure 5. Node monitoring in our platform.
Electronics 13 02651 g005
Figure 6. Easy deployment for any docker image in our Harbor registry.
Figure 6. Easy deployment for any docker image in our Harbor registry.
Electronics 13 02651 g006
Figure 7. Multiple options are available for ML analysis for workload placement.
Figure 7. Multiple options are available for ML analysis for workload placement.
Electronics 13 02651 g007
Figure 8. Discord chatbot provides information about workload metrics, starts workloads, stops workloads, etc.
Figure 8. Discord chatbot provides information about workload metrics, starts workloads, stops workloads, etc.
Electronics 13 02651 g008
Figure 9. Adding storage to our platform by using a built-in wizard.
Figure 9. Adding storage to our platform by using a built-in wizard.
Electronics 13 02651 g009
Figure 10. PDU monitoring in our platform.
Figure 10. PDU monitoring in our platform.
Electronics 13 02651 g010
Figure 11. Workload placement in Kubernetes when using our proposed platform.
Figure 11. Workload placement in Kubernetes when using our proposed platform.
Electronics 13 02651 g011
Figure 12. Manual workload placement in our platform.
Figure 12. Manual workload placement in our platform.
Electronics 13 02651 g012
Figure 13. Explanation of the scoring system that influences the ML decision-making process for workload scheduling.
Figure 13. Explanation of the scoring system that influences the ML decision-making process for workload scheduling.
Electronics 13 02651 g013
Table 1. Response times for our test scenario with a web application.
Table 1. Response times for our test scenario with a web application.
Test ScenarioEstimated Neural Network Response TimeCustom Scheduler Response TimeMean Response Time ErrorDefault Scheduler Response Time
No workloads placed6.455.890.565.94
One node is used for workloads6.46.090.316.15
Two nodes are used for workloads6.536.40.136.95
Three nodes are used for workloads6.66.690.097.88
Four nodes are used for workloads7.5810.763.1812.72
Table 2. Response times for our test scenario with HPCG-NVIDIA.
Table 2. Response times for our test scenario with HPCG-NVIDIA.
Test ScenarioEstimated Neural Network Response TimeCustom Scheduler Response TimeMean Response Time ErrorDefault Scheduler Response Time
No workloads placed9.329.080.249.28
One node is used for workloads10.139.760.3710.03
Two nodes are used for workloads12.4111.780.6312.13
Three nodes are used for workloads13.7813.370.4113.59
Four nodes are used for workloads14.9614.410.5514.88
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dakić, V.; Kovač, M.; Slovinac, J. Evolving High-Performance Computing Data Centers with Kubernetes, Performance Analysis, and Dynamic Workload Placement Based on Machine Learning Scheduling. Electronics 2024, 13, 2651. https://doi.org/10.3390/electronics13132651

AMA Style

Dakić V, Kovač M, Slovinac J. Evolving High-Performance Computing Data Centers with Kubernetes, Performance Analysis, and Dynamic Workload Placement Based on Machine Learning Scheduling. Electronics. 2024; 13(13):2651. https://doi.org/10.3390/electronics13132651

Chicago/Turabian Style

Dakić, Vedran, Mario Kovač, and Jurica Slovinac. 2024. "Evolving High-Performance Computing Data Centers with Kubernetes, Performance Analysis, and Dynamic Workload Placement Based on Machine Learning Scheduling" Electronics 13, no. 13: 2651. https://doi.org/10.3390/electronics13132651

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop