Low Latency Execution Guarantee Under Uncertainty in Serverless Platforms

Serverless computing recently emerged as a new run-time paradigm to disentangle the client from the burden of provisioning physical computing resources, leaving such difficulty on the service provider’s side. However, an unsolved problem in such an environment is how to cope with the challenges of e...

Full description

Saved in:
Bibliographic Details
Published inParallel and Distributed Computing, Applications and Technologies Vol. 13148; pp. 324 - 335
Main Authors HoseinyFarahabady, M. Reza, Taheri, Javid, Zomaya, Albert Y., Tari, Zahir
Format Book Chapter Conference Proceeding
LanguageEnglish
Published Switzerland Springer International Publishing AG 2022
Springer International Publishing
SeriesLecture Notes in Computer Science
Subjects
Online AccessGet full text
ISBN9783030967710
3030967719
3030967727
9783030967727
ISSN0302-9743
1611-3349
DOI10.1007/978-3-030-96772-7_30

Cover

More Information
Summary:Serverless computing recently emerged as a new run-time paradigm to disentangle the client from the burden of provisioning physical computing resources, leaving such difficulty on the service provider’s side. However, an unsolved problem in such an environment is how to cope with the challenges of executing several co-running applications while fulfilling the requested Quality of Service (QoS) level requested by all application owners. In practice, developing an efficient mechanism to reach the requested performance level (such as p-99 latency and throughput) is limited to the awareness (resource availability, performance interference among consolidation workloads, etc.) of the controller about the dynamics of the underlying platforms. In this paper, we develop an adaptive feedback controller for coping with the buffer instability of serverless platforms when several collocated applications are run in a shared environment. The goal is to support a low-latency execution by managing the arrival event rate of each application when shared resource contention causes a significant throughput degradation among workloads with different priorities. The key component of the proposed architecture is a continues management of server-side internal buffers for each application to provide a low-latency feedback control mechanism based on the requested QoS level of each application (e.g., buffer information) and the worker nodes throughput. The empirical results confirm the response stability for high priority workloads when a dynamic condition is caused by low priority applications. We evaluate the performance of the proposed solution with respect to the response time and the QoS violation rate for high priority applications in a serverless platform with four worker nodes set up in our in-house virtualized cluster. We compare the proposed architecture against the default resource management policy in Apache OpenWhisk which is extensively used in commercial serverless platforms. The results show that our approach achieves a very low overhead (less than 0.7%) while it can improve the p-99 latency of high priority applications by 64%, on average, in the presence of dynamic high traffic conditions.
ISBN:9783030967710
3030967719
3030967727
9783030967727
ISSN:0302-9743
1611-3349
DOI:10.1007/978-3-030-96772-7_30