umu.sePublications
Change search
Link to record
Permanent link

Direct link
BETA
Souza, Abel, PhLic.
Publications (4 of 4) Show all publications
Souza, A., Rezaei, M., Laure, E. & Tordsson, J. (2019). Hybrid Resource Management for HPC and Data Intensive Workloads. In: 2019 19th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID): . Paper presented at CCGrid2019, 14th - 17th of May Larnaca, Cyprus (pp. 399-409). Los Alamitos: IEEE Computer Society
Open this publication in new window or tab >>Hybrid Resource Management for HPC and Data Intensive Workloads
2019 (English)In: 2019 19th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGRID), Los Alamitos: IEEE Computer Society, 2019, p. 399-409Conference paper, Published paper (Refereed)
Abstract [en]

Traditionally, High Performance Computing (HPC) and Data Intensive (DI) workloads have been executed on separate hardware using different tools for resource and application management. With increasing convergence of these paradigms, where modern applications are composed of both types of jobs in complex workflows, this separation becomes a growing overhead and the need for a common computation platform for both application areas increases. Executing both application classes on the same hardware not only enables hybrid workflows, but can also increase the usage efficiency of the system, as often not all available hardware is fully utilized by an application. While HPC systems are typically managed in a coarse grained fashion, allocating a fixed set of resources exclusively to an application, DI systems employ a finer grained regime, enabling dynamic resource allocation and control based on application needs. On the path to full convergence, a useful and less intrusive step is a hybrid resource management system that allows the execution of DI applications on top of standard HPC scheduling systems.In this paper we present the architecture of a hybrid system enabling dual-level scheduling for DI jobs in HPC infrastructures. Our system takes advantage of real-time resource utilization monitoring to efficiently co-schedule HPC and DI applications. The architecture is easily adaptable and extensible to current and new types of distributed workloads, allowing efficient combination of hybrid workloads on HPC resources with increased job throughput and higher overall resource utilization. The architecture is implemented based on the Slurm and Mesos resource managers for HPC and DI jobs. Our experimental evaluation in a real cluster based on a set of representative HPC and DI applications demonstrate that our hybrid architecture improves resource utilization by 20%, with 12% decrease on queue makespan while still meeting all deadlines for HPC jobs.

Place, publisher, year, edition, pages
Los Alamitos: IEEE Computer Society, 2019
Keywords
Resource Management, High Performance Computing, Data Intensive Computing, Mesos, Slurm, Boostrapping
National Category
Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:umu:diva-155619 (URN)10.1109/CCGRID.2019.00054 (DOI)978-1-7281-0913-8 (ISBN)978-1-7281-0912-1 (ISBN)
Conference
CCGrid2019, 14th - 17th of May Larnaca, Cyprus
Note

Originally included in thesis in manuscript form

Available from: 2019-01-24 Created: 2019-01-24 Last updated: 2019-11-18Bibliographically approved
Souza, A. P. (2018). Application-aware resource management for datacenters. (Licentiate dissertation). Umeå: Department of computing science, Umeå university
Open this publication in new window or tab >>Application-aware resource management for datacenters
2018 (English)Licentiate thesis, comprehensive summary (Other academic)
Alternative title[sv]
Applikationsmedveten resurshantering för datacenter
Abstract [en]

High Performance Computing (HPC) and Cloud Computing datacenters are extensively used to steer and solve complex problems in science, engineering, and business, such as calculating correlations and making predictions. Already in a single datacenter server, there are thousands of hardware and software metrics – Key Performance Indicators (KPIs) – that individually and aggregated can give insight in the performance, robustness, and efficiency of the datacenter and the provisioned applications. At the datacenter level, the number of KPIs is even higher. The fast growing interest on datacenter management from both public and industry together with the rapid expansion in scale and complexity of datacenter resources and the services being provided on them have made monitoring, profiling, controlling, and provisioning compute resources dynamically at runtime into a challenging and complex task. Commonly, correlations of application KPIs, like response time and throughput, with resource capacities show that runtime systems (e.g., containers or virtual machines) that are used to provision these applications do not utilize available resources efficiently. This reduces datacenter efficiency, which in term results in higher operational costs and longer waiting times for results.

The goal of this thesis is to develop tools and autonomic techniques for improving datacenter operations, management and utilization, while improving and/or minimizing impacts on applications performance. To this end, we make use of application resource descriptors to create a library that dynamically adjusts the amount of resources used, enabling elasticity for scientific workflows in HPC datacenters. For mission critical applications, high availability is of great concern since these services must be kept running even in the event of system failures. By modeling and correlating specific resource counters, like CPU, memory and network utilization, with the number of runtime synchronizations, we present adaptive mechanisms to dynamically select which fault tolerant mechanism to use. Likewise, for scientific applications we propose a hybrid extensible architecture for dual-level scheduling of data intensive jobs in HPC infrastructures, allowing operational simplification, on-boarding of new types of applications and achieving greater job throughput with higher overall datacenter efficiency.

Place, publisher, year, edition, pages
Umeå: Department of computing science, Umeå university, 2018. p. 28
Series
Report / UMINF, ISSN 0348-0542 ; 18.14
Keywords
Resource Management, High Performance Computing, Cloud Computing
National Category
Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:umu:diva-155620 (URN)978-91-7601-971-9 (ISBN)
Presentation
2018-12-12, MA121, MIT-Huset, Umeå, 20:31 (English)
Opponent
Supervisors
Available from: 2019-01-25 Created: 2019-01-24 Last updated: 2019-02-04Bibliographically approved
Souza, A., Papadopoulos, A. V., Tomás Bolivar, L., Gilbert, D. & Tordsson, J. (2018). Hybrid Adaptive Checkpointing for Virtual Machine Fault Tolerance. In: Li J., Chandra A., Guo T., Cai Y. (Ed.), Proceedings - 2018 IEEE International Conference on Cloud Engineering, IC2E 2018: . Paper presented at 2018 IEEE International Conference on Cloud Engineering (IC2E 2018), 17–20 April 2018, Orlando, Florida, USA (pp. 12-22). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Hybrid Adaptive Checkpointing for Virtual Machine Fault Tolerance
Show others...
2018 (English)In: Proceedings - 2018 IEEE International Conference on Cloud Engineering, IC2E 2018 / [ed] Li J., Chandra A., Guo T., Cai Y., Institute of Electrical and Electronics Engineers (IEEE), 2018, p. 12-22Conference paper, Published paper (Refereed)
Abstract [en]

Active Virtual Machine (VM) replication is an application independent and cost-efficient mechanism for high availability and fault tolerance, with several recently proposed implementations based on checkpointing. However, these methods may suffer from large impacts on application latency, excessive resource usage overheads, and/or unpredictable behavior for varying workloads. To address these problems, we propose a hybrid approach through a Proportional-Integral (PI) controller to dynamically switch between periodic and on-demand check-pointing. Our mechanism automatically selects the method that minimizes application downtime by adapting itself to changes in workload characteristics. The implementation is based on modifications to QEMU, LibVirt, and OpenStack, to seamlessly provide fault tolerant VM provisioning and to enable the controller to dynamically select the best checkpointing mode. Our evaluation is based on experiments with a video streaming application, an e-commerce benchmark, and a software development tool. The experiments demonstrate that our adaptive hybrid approach improves both application availability and resource usage compared to static selection of a checkpointing method, with application performance gains and neglectable overheads.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2018
Keywords
Fault Tolerance, Resource Management, Checkpoint, COLO, Control Theory
National Category
Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:umu:diva-152033 (URN)10.1109/IC2E.2018.00023 (DOI)2-s2.0-85048315473 (Scopus ID)978-1-5386-5009-7 (ISBN)978-1-5386-5008-0 (ISBN)
Conference
2018 IEEE International Conference on Cloud Engineering (IC2E 2018), 17–20 April 2018, Orlando, Florida, USA
Available from: 2018-09-24 Created: 2018-09-24 Last updated: 2019-01-24Bibliographically approved
Fox, W., Ghoshal, D., Souza, A., P. Rodrigo, G. & Ramakrishnan, L. (2017). E-HPC: A Library for Elastic Resource Management in HPC Environments. In: 12th Workshop on Workflows in Support of Large-Scale Science (WORKS): . Paper presented at The International Conference for High Performance Computing, Networking, Storage and Analysis. New York, NY, USA: Association for Computing Machinery (ACM), Article ID 1.
Open this publication in new window or tab >>E-HPC: A Library for Elastic Resource Management in HPC Environments
Show others...
2017 (English)In: 12th Workshop on Workflows in Support of Large-Scale Science (WORKS), New York, NY, USA: Association for Computing Machinery (ACM), 2017, article id 1Conference paper, Published paper (Refereed)
Abstract [en]

Next-generation data-intensive scientific workflows need to support streaming and real-time applications with dynamic resource needs on high performance computing (HPC) platforms. The static resource allocation model on current HPC systems that was designed for monolithic MPI applications is insufficient to support the elastic resource needs of current and future workflows. In this paper, we discuss the design, implementation and evaluation of Elastic-HPC (E-HPC), an elastic framework for managing resources for scientific workflows on current HPC systems. E-HPC considers a resource slot for a workflow as an elastic window that might map to different physical resources over the duration of a workflow. Our framework uses checkpoint-restart as the underlying mechanism to migrate workflow execution across the dynamic window of resources. E-HPC provides the foundation necessary to enable dynamic resource allocation of HPC resources that are needed for streaming and real-time workflows. E-HPC has negligible overhead beyond the cost of checkpointing. Additionally, E-HPC results in decreased turnaround time of workflows compared to traditional model of resource allocation for workflows, where resources are allocated per stage of the workflow. Our evaluation shows that E-HPC improves core hour utilization for common workflow resource use patterns and provides an effective framework for elastic expansion of resources for applications with dynamic resource needs.

Place, publisher, year, edition, pages
New York, NY, USA: Association for Computing Machinery (ACM), 2017
Keywords
high performance computing, scientific workflows, resource management
National Category
Computer Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:umu:diva-142624 (URN)10.1145/3150994.3150996 (DOI)978-1-4503-5129-4 (ISBN)
Conference
The International Conference for High Performance Computing, Networking, Storage and Analysis
Available from: 2017-12-06 Created: 2017-12-06 Last updated: 2019-01-24Bibliographically approved
Organisations

Search in DiVA

Show all publications