Is it time to put cold starts in the deep freeze?
File(s)deep_freeze-socc24.pdf (912.46 KB)
Accepted version
Author(s)
Segarra, Carlos
Durev, Ivan
Pietzuch, Peter
Type
Conference Paper
Abstract
Cold-start times have been the "end-all, be-all" metric for research in serverless cloud computing over the past decade. Reducing the impact of cold starts matters, because they can be the biggest contributor to a serverless function's end-to-end execution time. Recent studies from cloud providers, however, indicate that, in practice, a majority of serverless functions are triggered by non-interactive workloads. To substantiate this, we study the types of serverless functions used in 35 publications and find that over 80% of functions are not semantically latency sensitive. If a function is non-interactive and latency insensitive, is end-to-end execution time the right metric to optimize in serverless? What if cold starts do not matter that much, after all?
In this vision paper, we explore what serverless environments in which cold starts do not matter would look like. We make the case that serverless research should focus on supporting latency insensitive, i.e., batch, workloads. Based on this, we explore the design space for DFaaS, a serverless framework with an execution model in which functions can be arbitrarily delayed. DFaaS users annotate each function with a delay tolerance and, as long as the deadline has not passed, the runtime may interrupt or migrate function execution. Our micro-benchmarks suggest that, by targeting batch workloads, DFaaS can improve substantially the resource usage of serverless clouds and lower costs for users.
In this vision paper, we explore what serverless environments in which cold starts do not matter would look like. We make the case that serverless research should focus on supporting latency insensitive, i.e., batch, workloads. Based on this, we explore the design space for DFaaS, a serverless framework with an execution model in which functions can be arbitrarily delayed. DFaaS users annotate each function with a delay tolerance and, as long as the deadline has not passed, the runtime may interrupt or migrate function execution. Our micro-benchmarks suggest that, by targeting batch workloads, DFaaS can improve substantially the resource usage of serverless clouds and lower costs for users.
Date Issued
2024-11-01
Date Acceptance
2024-11-01
Citation
Proceedings of the ACM Symposium on Cloud Computing, 2024, pp.259-268
ISBN
9798400712869
Publisher
ACM
Start Page
259
End Page
268
Journal / Book Title
Proceedings of the ACM Symposium on Cloud Computing
Copyright Statement
© 2024 Copyright held by the owner/author(s). Publication rights licensed to ACM. This is the author’s accepted manuscript made available under a CC-BY licence in accordance with Imperial’s Research Publications Open Access policy (www.imperial.ac.uk/oa-policy)
License URL
Source
SoCC '24: ACM Symposium on Cloud Computing
Publication Status
Published
Start Date
2024-11-20
Finish Date
2024-11-22
Coverage Spatial
Redmond, WA, USA
Date Publish Online
2024-11-20