Taming cold starts: proactive serverless scheduling with model predictive control
2025 (Engelska)Konferensbidrag, Enbart muntlig presentation (Refereegranskat)
Abstract [en]
Serverless computing has transformed cloud application deployment by introducing a fine-grained, event-driven execution model that abstracts away infrastructure management. Its on-demand nature makes it especially appealing for latency-sensitive and bursty workloads. However, the cold start problem, i.e., where the platform incurs significant delay when provisioning new containers, remains the Achilles' heel of such platforms.
This paper presents a predictive serverless scheduling framework based on Model Predictive Control to proactively mitigate cold starts, thereby improving end-to-end response time. By forecasting future invocations, the controller jointly optimizes container prewarming and request dispatching, improving latency while minimizing resource overhead.
We implement our approach on Apache OpenWhisk, deployed on a Kubernetes-based testbed. Experimental results using real-world function traces and synthetic workloads demonstrate that our method significantly outperforms state-of-the-art baselines, achieving up to 85% lower tail latency and a 34% reduction in resource usage.
Ort, förlag, år, upplaga, sidor
2025.
Nyckelord [en]
Serverless, Cloud Computing, Orchestration, Cold Start, Function-as-a-service, Model Predictive Control, Prediction, Request Shaping
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
datalogi
Identifikatorer
URN: urn:nbn:se:umu:diva-243593OAI: oai:DiVA.org:umu-243593DiVA, id: diva2:1994897
Konferens
MASCOTS 2025: 33rd International Symposium on the Modeling, Analysis, and Simulation of Computer and Telecommunication System, Paris, France, October 21-23, 2025
Forskningsfinansiär
Wallenberg AI, Autonomous Systems and Software Program (WASP)EU, Horisont Europa, 1010927112025-09-032025-09-032025-09-04