Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Location-aware resource allocation in mobile edge clouds
Umeå University, Faculty of Science and Technology, Department of Computing Science. (Autonomous Distributed Systems Lab)ORCID iD: 0000-0002-9156-3364
2021 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Over the last decade, cloud computing has realized the long-held dream of computing as a utility, in which computational and storage services are made available via the Internet to anyone at any time and from anywhere. This has transformed Information Technology (IT) and given rise to new ways of designing and purchasing hardware and software. However, the rapid development of the Internet of Things (IoTs) and mobile technology has brought a new wave of disruptive applications and services whose performance requirements are stretching the limits of current cloud computing systems and platforms. In particular, novel large scale mission-critical IoT systems and latency-intolerant applications strictly require very low latency and strong guarantees of privacy, and can generate massive amounts of data that are only of local interest. These requirements are not readily satisfied using modern application deployment strategies that rely on resources from distant large cloud datacenters because they easily cause network congestion and high latency in service delivery. This has provoked a paradigm shift leading to the emergence of new distributed computing infrastructures known as Mobile Edge Clouds (MECs) in which resource capabilities are widely distributed at the edge of the network, in close proximity to end-users.  Experimental studies have validated and quantified many benefits of MECs, which include considerable improvements in response times and enormous reductions in ingress bandwidth demand. However, MECs must cope with several challenges not commonly encountered in traditional cloud systems, including user mobility, hardware heterogeneity, and considerable flexibility in terms of where computing capacity can be used. This makes it especially difficult to analyze, predict, and control resource usage and allocation so as to minimize cost and maximize performance while delivering the expected end-user Quality-of-Service (QoS). Realizing the potential of MECs will thus require the design and development of efficient resource allocation systems that take these factors into consideration. 

Since the introduction of the MEC concept, the performance benefits achieved by running MEC-native applications (i.e., applications engineered specifically for MECs) on MECs have been clearly demonstrated. However, the benefits of MECs for non-MEC-native applications (i.e., application not specifically engineered for MECs) are still questioned. This is a fundamental issue that must be explored because it will affect the incentives for service providers and application developers to invest in MECs. To spur the development of MECs, the first part of this thesis presents an extensive investigation of the benefits that MECs can offer to non-MEC-native applications. One class of non-MEC-native applications that could potentially benefit significantly from deployment on an MEC is cloud-native applications, particularly micro-service-based applications with high deployment flexibility. We therefore quantitatively compared the performance of cloud-native applications deployed using resources from cloud datacenters and edge locations. We then developed a network communication profiling tool to identify aspects of these applications that reduce the benefits derived from deployment on MECs, and proposed design improvements that would allow such applications to better exploit MECs' capabilities.  

The second part of this thesis addresses problems related to resource allocation in highly distributed MECs. First, to overcome challenges arising from the dynamic nature of resource demand in MECs, we used statistical time series models and machine learning techniques to develop two location-aware workload prediction models for EDCs that account for both user mobility and the correlation of workload changes among EDCs in close physical proximity. These models were then utilized to develop an elasticity controller for MECs. In essence, the controller helps MECs to perform resource allocation, i.e. to answer the intertwined questions of what and how many resources should be allocated and when and where they should be deployed.

The third part of the thesis focuses on problems relating to the real-time placement of stateful applications on MECs. Specifically, it examines the questions of where to place applications so as to minimize total operating costs while delivering the required end-user QoS and whether the requested applications should be migrated to follow the user's movements. Such questions are easy to pose but intrinsically hard to answer due to the scale and complexity of MEC infrastructures and the stochastic nature of user mobility. To this end, we first thoroughly modeled the workloads, stateful applications, and infrastructures to be expected in MECs. We then formulated the various costs associated with operating applications, namely the resource cost, migration cost, and service quality degradation cost. Based on our model, we proposed two online application placement algorithms that take these factors into account to minimize the total cost of operating the application.

The methods and algorithms proposed in this thesis were evaluated by implementing prototypes on simulated testbeds and conducting experiments using workloads based on real mobility traces. These evaluations showed that the proposed approaches outperformed alternative state-of-the-art approaches and could thus help improve the efficiency of resource allocation in MECs.

Place, publisher, year, edition, pages
Umeå: Umeå universitet , 2021. , p. 50
Series
Report / UMINF, ISSN 0348-0542 ; 21.01
Keywords [en]
Mobile Edge Clouds, Resource Allocation, Quality of Service, Application Placement, Workload Prediction
National Category
Computer Sciences
Research subject
Computer Systems; Computer Science
Identifiers
URN: urn:nbn:se:umu:diva-178926ISBN: 978-91-7855-467-6 (print)ISBN: 978-91-7855-466-9 (electronic)OAI: oai:DiVA.org:umu-178926DiVA, id: diva2:1520686
Public defence
2021-02-17, Aula Biologica, Umeå Universitet, 901 87 Umeå, Umeå, 13:00 (English)
Opponent
Supervisors
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)Available from: 2021-01-27 Created: 2021-01-21 Last updated: 2021-06-11Bibliographically approved
List of papers
1. Why Cloud Applications Are not Ready for the Edge (yet)
Open this publication in new window or tab >>Why Cloud Applications Are not Ready for the Edge (yet)
2019 (English)In: Proceedings of the 4th ACM/IEEE Symposium on Edge Computing, IEEE, 2019, p. 250-263Conference paper, Published paper (Other academic)
Abstract [en]

Mobile Edge Clouds (MECs) are distributed platforms in which distant data-centers are complemented with computing and storage capacity located at the edge of the network. Their wide resource distribution enables MECs to fulfill the need of low latency and high bandwidth to offer an improved user experience.

As modern cloud applications are increasingly architected as collections of small, independently deployable services, they can be flexibly deployed in various configurations that combines resources from both centralized datacenters and edge locations. In principle, such applications should therefore be well-placed to exploit the advantages of MECs so as to reduce service response times.

In this paper, we quantify the benefits of deploying such cloud micro-service applications on MECs. Using two popular benchmarks, we show that, against conventional wisdom, end-to-end latency does not improve significantly even when most application services are deployed in the edge location. We developed a profiler to better understand this phenomenon, allowing us to develop recommendations for adapting applications to MECs. Further, by quantifying the gains of those recommendations, we show that the performance of an application can be made to reach the ideal scenario, in which the latency between an edge datacenter and a remote datacenter has no impact on the application performance.

This work thus presents ways of adapting cloud-native applications to take advantage of MECs and provides guidance for developing MEC-native applications. We believe that both these elements are necessary to drive MEC adoption.

Place, publisher, year, edition, pages
IEEE, 2019
Keywords
Mobile Edge Clouds, Edge Latency, Mobile Application Development, Micro-service, Profiling
National Category
Computer Sciences
Identifiers
urn:nbn:se:umu:diva-162930 (URN)10.1145/3318216.3363298 (DOI)000680020300019 ()2-s2.0-85076258710 (Scopus ID)978-1-4503-6733-2 (ISBN)
Conference
The Fourth ACM/IEEE Symposium on Edge Computing, Washington DC, November 7–9, 2019
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Available from: 2019-09-02 Created: 2019-09-02 Last updated: 2023-09-05Bibliographically approved
2. Location-aware load prediction in edge data centers
Open this publication in new window or tab >>Location-aware load prediction in edge data centers
2017 (English)In: 2017 Second International Conference on Fog and Mobile Edge Computing (FMEC), IEEE, 2017, p. 25-31Conference paper, Published paper (Other academic)
Abstract [en]

Mobile Edge Cloud (MEC) is a platform complementing traditional centralized clouds, consisting in moving computing and storage capacity closer to users -e. g., as Edge Data Centers (EDC) in base stations -in order to reduce application-level latency and network bandwidth. The bounded coverage radius of base station and limited capacity of each EDC intertwined with user mobility challenge the operator's ability to perform capacity adjustment and planning. To face this challenge, proactive resource provisioning can be performed. The resource usage in each EDC is estimated in advance, which is made available for the decision making to efficiently determine various management actions and ensure that EDCs persistently satisfies the Quality of Service (QoS), while maximizing resource utilization. In this paper, we propose location-aware load prediction. For each EDC, load is not only predicted using its own historical load time series -as done for centralized clouds -but also those of its neighbor EDCs. We employ Vector Autoregression Model (VAR) in which the correlation among adjacent EDCs load time series are exploited. We evaluate our approach using real world mobility traces to simulate load in each EDC and conduct various experiments to evaluate the proposed algorithm. Result shows that our proposed algorithm is able to achieve an average accuracy of up to 93% on EDCs with substantial average load, which slightly improves prediction by 4.3% compared to the state-of-the-art approach. Considering the expected scale of MEC, this translates to substantial cost savings e. g., servers can be shutdown without QoS violation.

Place, publisher, year, edition, pages
IEEE, 2017
Keywords
Workload Prediction, Proactive Resource Management, Mobile Edge Cloud, VAR Model, User Mobility
National Category
Computer Systems
Identifiers
urn:nbn:se:umu:diva-135700 (URN)000411731700004 ()978-1-5386-2859-1 (ISBN)
Conference
The 2nd International Conference on Fog and Mobile Edge Computing (FMEC), May 8-11, 2017, Valencia, Spain
Projects
WASP
Available from: 2017-06-02 Created: 2017-06-02 Last updated: 2021-03-18Bibliographically approved
3. Multivariate LSTM-Based Location-Aware Workload Prediction for Edge Data Centers
Open this publication in new window or tab >>Multivariate LSTM-Based Location-Aware Workload Prediction for Edge Data Centers
2019 (English)In: Proceedings, 19th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing, IEEE, 2019, p. 341-350Conference paper, Published paper (Refereed)
Abstract [en]

Mobile Edge Clouds (MECs) is a promising computing platform to overcome challenges for the success of bandwidth-hungry, latency-critical applications by distributing computing and storage capacity in the edge of the network as Edge Data Centers (EDCs) within the close vicinity of end-users. Due to the heterogeneous distributed resource capacity in EDCs, the application deployment flexibility coupled with the user mobility, MECs bring significant challenges to control resource allocation and provisioning. In order to develop a self-managed system for MECs which efficiently decides how much and when to activate scaling, where to place and migrate services, it is crucial to predict its workload characteristics, including variations over time and locality. To this end, we present a novel location-aware workload predictor for EDCs. Our approach leverages the correlation among workloads of EDCs in a close physical distance and applies multivariate Long Short-Term Memory network to achieve on-line workload predictions for each EDC. The experiments with two real mobility traces show that our proposed approach can achieve better prediction accuracy than a state-of-the art location-unaware method (up to 44%) and a location-aware method (up to 17%). Further, through an intensive performance measurement using various input shaking methods, we substantiate that the proposed approach achieves a reliable and consistent performance.

Place, publisher, year, edition, pages
IEEE, 2019
Keywords
Mobile Edge Cloud, Edge Data Center, ResourceManagement, Workload Prediction, Location-aware, MachineLearning
National Category
Computer Systems
Research subject
Computer Systems
Identifiers
urn:nbn:se:umu:diva-159540 (URN)10.1109/CCGRID.2019.00048 (DOI)000483058700039 ()2-s2.0-85069517887 (Scopus ID)978-1-7281-0912-1 (ISBN)978-1-7281-0913-8 (ISBN)
Conference
CCGrid 2019, 19th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (IEEE/ACM CCGrid 2019), 14-17 May, Larnaca, Cyprus
Available from: 2019-05-30 Created: 2019-05-30 Last updated: 2021-03-18Bibliographically approved
4. Elasticity Control for Latency-IntolerantMobile Edge Applications
Open this publication in new window or tab >>Elasticity Control for Latency-IntolerantMobile Edge Applications
2020 (English)In: The Fifth ACM/IEEE Symposium on Edge Computing, Virtual, November 11-13, 2020, Institute of Electrical and Electronics Engineers (IEEE), 2020, p. 70-83Conference paper, Published paper (Refereed)
Abstract [en]

Elasticity is a fundamental property required for Mobile Edge Clouds (MECs) to become mature computing platforms hosting software applications. However, MECs must cope with several challenges that do not arise in the context of conventional cloud platforms. These include the potentially highly distributed geographical deployment, heterogeneity, and limited resource capacity of Edge Data Centers (EDCs), and end-user mobility.

In this paper, we present an elasticity controller to help MECs overcome these challenges by automatic proactive resource scaling. The controller utilizes information on the physical locations of EDCs and the correlation of workload changes in physically neighboring EDCs to predict request arrival rates at EDCs. These predictions are used as inputs for a queueing theory-driven performance model that estimates the number of resources that should be provisioned to EDCs in order to meet predefined Service Level Objectives (SLOs) while maximizing resource utilization. The controller also incorporates a group-level load balancer that is responsible for redirecting requests among EDCs during runtime so as to minimize the request rejection rate.

We evaluate our approach by performing simulations with an emulated MEC deployed over a metropolitan area and a simulated application workload using a real-world user mobility trace. The results show that our proposed pro-active controller exhibits better scaling behavior than a state-of-the-art re-active controller and increases the efficiency of resource provisioning, thereby helping MECs to sustain resource utilization and rejection rates that satisfy predefined SLOs while maintaining system stability.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2020
Keywords
Resource Provisioning, Elasticity, Auto Scaling, Edge Data Center, Workload Prediction, Location-aware, Machine Learning.
National Category
Computer Systems
Research subject
Computer Systems; Computer Science
Identifiers
urn:nbn:se:umu:diva-177976 (URN)10.1109/SEC50012.2020.00013 (DOI)000667978500006 ()2-s2.0-85102183446 (Scopus ID)
Conference
SEC 2020, The Fifth ACM/IEEE Symposium on Edge Computing, Virtual, November 11-13, 2020
Available from: 2020-12-28 Created: 2020-12-28 Last updated: 2023-09-05Bibliographically approved
5. State-aware application placement in mobile edge clouds
Open this publication in new window or tab >>State-aware application placement in mobile edge clouds
2024 (English)In: Proceedings of the 14th international conference on cloud computing and services science / [ed] Maarten van Steen; Claus Pahl, Portugal: Science and Technology Publications , 2024, Vol. 1, p. 117-128Conference paper, Published paper (Refereed)
Abstract [en]

Placing applications within Mobile Edge Clouds (MEC) poses challenges due to dynamic user mobility. Maintaining optimal Quality of Service may require frequent application migration in response to changing user locations, potentially leading to bandwidth wastage. This paper addresses application placement challenges in MEC environments by developing a comprehensive model covering workloads, applications, and MEC infrastructures. Following this, various costs associated with application operation, including resource utilization, migration overhead, and potential service quality degradation, are systematically formulated. An online application placement algorithm, App EDC Match, inspired by the Gale-Shapley matching algorithm, is introduced to optimize application placement considering these cost factors. Through experiments that employ real mobility traces to simulate workload dynamics, the results demonstrate that the proposed algorithm efficiently determines near-optimal application placements within Edge Data Centers. It achieves total operating costs within a narrow margin of 8% higher than the approximate global optimum attained by the offline precognition algorithm, which assumes access to future user locations. Additionally, the proposed placement algorithm effectively mitigates resource scarcity in MEC.

Place, publisher, year, edition, pages
Portugal: Science and Technology Publications, 2024
Series
International Conference on Cloud Computing and Services Science, E-ISSN 2184-5042
Keywords
Mobile Edge Clouds, Application Placement, Service Orchestration, Optimization
National Category
Computer Sciences Computer Systems
Research subject
Computer Science; Computer Systems
Identifiers
urn:nbn:se:umu:diva-178830 (URN)10.5220/0012326300003711 (DOI)2-s2.0-85194153990 (Scopus ID)9789897587016 (ISBN)
Conference
14th International Conference on Cloud Computing and Services Science, CLOSER 2024, Angers, France, May 2-4, 2024
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

Originally included in thesis in manuscript form.

Available from: 2021-01-19 Created: 2021-01-19 Last updated: 2024-07-17Bibliographically approved

Open Access in DiVA

fulltext(1198 kB)556 downloads
File information
File name FULLTEXT01.pdfFile size 1198 kBChecksum SHA-512
062eb1417df7c884f434a55a112a01adf5e906df29e77b621216ec89733a4ba88a8864ff41197360521b895088604bcdd5389858b240e243f8a6c0f763b431a4
Type fulltextMimetype application/pdf
spikblad(153 kB)139 downloads
File information
File name SPIKBLAD01.pdfFile size 153 kBChecksum SHA-512
a1e735c913399366c89fe8151dba58e7ba467fe6680485e17d15921b1e29d247680db96e712bcd8a4a26820018662cf8b1a35a844fbdc4a1cbdc2056b57b01eb
Type spikbladMimetype application/pdf
cover(3262 kB)104 downloads
File information
File name COVER01.pdfFile size 3262 kBChecksum SHA-512
49bace88854081890cc719d49e222cb42b973a18f0677dbcb35708f547da792b256d0f7b1b7bb3ec73d8f1fa72163f13661dcc0b3ae569401fc5ac7d6c6f9923
Type coverMimetype application/pdf

Authority records

Nguyen, Chanh Le Tan

Search in DiVA

By author/editor
Nguyen, Chanh Le Tan
By organisation
Department of Computing Science
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 556 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 1920 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf