TY - GEN
T1 - Towards Virtualization-Agnostic Latency for Time-Sensitive Applications
AU - Li, Haoran
AU - Xu, Meng
AU - Li, Chong
AU - Lu, Chenyang
AU - Gill, Christopher
AU - Phan, Linh
AU - Lee, Insup
AU - Sokolsky, Oleg
N1 - Publisher Copyright:
© 2021 ACM.
PY - 2021/4/7
Y1 - 2021/4/7
N2 - As time-sensitive applications are deployed spanning multiple edge clouds, delivering consistent and scalable latency performance across different virtualized hosts becomes increasingly challenging. In contrast to traditional real-time systems requiring deadline guarantees for all jobs, the latency service-level objectives of cloud applications are usually defined in terms of tail latency, i.e., the latency of a certain percentage of the jobs should be below a given threshold. This means that neither dedicating entire physical CPU cores, nor combining virtualization with deadline-based techniques such as compositional real-time scheduling, can meet the needs of these applications in a resource-efficient manner. To address this limitation, and to simplify the management of edge clouds for latency-sensitive applications, we introduce virtualization-agnostic latency (VAL) as an essential property to maintain consistent tail latency assurances across different virtualized hosts. VAL requires that an application experience similar latency distributions on a shared host as on a dedicated one. Towards achieving VAL in edge clouds, this paper presents a virtualization-agnostic scheduling (VAS) framework for time-sensitive applications sharing CPUs with other applications. We show both theoretically and experimentally that VAS can effectively deliver VAL on shared hosts. For periodic and sporadic tasks, we establish theoretical guarantees that VAS can achieve the same task schedule on a shared CPU as on a full CPU dedicated to time-sensitive services. Moreover, this can be achieved by allocating the minimal CPU bandwidth to time-sensitive services, thereby avoiding wasting CPU resources. VAS has been implemented on Xen 4.10.0. In case studies running time-sensitive workloads on Redis and Spark streaming services, we show that in practice the task schedule on a shared CPU can closely approximate the one on a full CPU.
AB - As time-sensitive applications are deployed spanning multiple edge clouds, delivering consistent and scalable latency performance across different virtualized hosts becomes increasingly challenging. In contrast to traditional real-time systems requiring deadline guarantees for all jobs, the latency service-level objectives of cloud applications are usually defined in terms of tail latency, i.e., the latency of a certain percentage of the jobs should be below a given threshold. This means that neither dedicating entire physical CPU cores, nor combining virtualization with deadline-based techniques such as compositional real-time scheduling, can meet the needs of these applications in a resource-efficient manner. To address this limitation, and to simplify the management of edge clouds for latency-sensitive applications, we introduce virtualization-agnostic latency (VAL) as an essential property to maintain consistent tail latency assurances across different virtualized hosts. VAL requires that an application experience similar latency distributions on a shared host as on a dedicated one. Towards achieving VAL in edge clouds, this paper presents a virtualization-agnostic scheduling (VAS) framework for time-sensitive applications sharing CPUs with other applications. We show both theoretically and experimentally that VAS can effectively deliver VAL on shared hosts. For periodic and sporadic tasks, we establish theoretical guarantees that VAS can achieve the same task schedule on a shared CPU as on a full CPU dedicated to time-sensitive services. Moreover, this can be achieved by allocating the minimal CPU bandwidth to time-sensitive services, thereby avoiding wasting CPU resources. VAS has been implemented on Xen 4.10.0. In case studies running time-sensitive workloads on Redis and Spark streaming services, we show that in practice the task schedule on a shared CPU can closely approximate the one on a full CPU.
KW - Deferrable Server
KW - Real-Time Scheduling
KW - Virtualization
UR - https://www.scopus.com/pages/publications/85111967025
U2 - 10.1145/3453417.3453420
DO - 10.1145/3453417.3453420
M3 - Conference contribution
AN - SCOPUS:85111967025
T3 - ACM International Conference Proceeding Series
SP - 35
EP - 45
BT - RTNS 2021 - 29th International Conference on Real-Time Networks and Systems
PB - Association for Computing Machinery
T2 - 29th International Conference on Real-Time Networks and Systems, RTNS 2021
Y2 - 7 April 2021
ER -