Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Increasing Efficiency and Result Reliability of Continuous Benchmarking for FaaS Applications (2405.15610v2)

Published 24 May 2024 in cs.DC

Abstract: In a continuous deployment setting, Function-as-a-Service (FaaS) applications frequently receive updated releases, each of which can cause a performance regression. While continuous benchmarking, i.e., comparing benchmark results of the updated and the previous version, can detect such regressions, performance variability of FaaS platforms necessitates thousands of function calls, thus, making continuous benchmarking time-intensive and expensive. In this paper, we propose DuetFaaS, an approach which adapts duet benchmarking to FaaS applications. With DuetFaaS, we deploy two versions of FaaS function in a single cloud function instance and execute them in parallel to reduce the impact of platform variability. We evaluate our approach against state-of-the-art approaches, running on AWS Lambda. Overall, DuetFaaS requires fewer invocations to accurately detect performance regressions than other state-of-the-art approaches. In 98.41% of evaluated cases, our approach provides equal or smaller confidence interval size. DuetFaaS achieves an interval size reduction in 59.06% of all evaluated sample sizes when compared to the competitive approaches.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. Ali Abedi and Tim Brecht. 2017. Conducting repeatable experiments in highly variable cloud computing environments. In Proceedings of the 8th ACM/SPEC on International Conference on Performance Engineering (ICPE ’17). 287–292.
  2. David Bermbach. 2017a. Quality of cloud services: Expect the unexpected. IEEE Internet Computing 21, 1 (2017), 68–72.
  3. David Bermbach. 2017b. Quality of Cloud Services: Expect the Unexpected. IEEE Internet Computing (Invited Paper) 21, 1 (2017), 68–72.
  4. On the future of cloud engineering. In Proceedings of the 2021 IEEE International conference on cloud engineering (IC2E ’21). 264–275.
  5. On the Future of Cloud Engineering. In Proceedings of the 9th IEEE International Conference on Cloud Engineering (IC2E ’21). 264–275.
  6. Using Application Knowledge to Reduce Cold Starts in FaaS Services. In Proceedings of the 35th ACM Symposium on Applied Computing (SAC ’20). 134–143.
  7. Cloud Service Benchmarking: Measuring Quality of Cloud Services from a Client Perspective. Springer.
  8. Initial experiments with duet benchmarking: Performance testing interference in the cloud. In Proceedings of the 2019 IEEE 27th International Symposium on Modeling, Analysis, and Simulation of Computer and Telecommunication Systems (MASCOTS ’19). 249–255.
  9. Duet benchmarking: Improving measurement accuracy in the cloud. In Proceedings of the ACM/SPEC International Conference on Performance Engineering (ICPE ’20). 100–107.
  10. Sebs: A serverless benchmark suite for function-as-a-service computing. In Proceedings of the 22nd International Middleware Conference (Middleware ’21). 64–78.
  11. The use of change point detection to identify software performance regressions in a continuous integration system. In Proceedings of the ACM/SPEC International Conference on Performance Engineering (ICPE ’20). 67–75.
  12. Efficiently Detecting Performance Changes in FaaS Application Releases. In Proceedings of the 9th International Workshop on Serverless Computing (WoSC ’23). 13–17.
  13. Using Microbenchmark Suites to Detect Application Performance Changes. IEEE Transactions on Cloud Computing 11, 3 (2023), 2575–2590.
  14. Continuous benchmarking: Using system benchmarking in build pipelines. In Proceedings of the 2019 IEEE International Conference on Cloud Engineering (IC2E ’19). 241–246.
  15. Befaas: An application-centric benchmarking framework for faas platforms. In Proceedings of the 2021 IEEE International Conference on Cloud Engineering (IC2E ’21). 1–8.
  16. Serverless computation with OpenLambda. In Proceedings of the 8th USENIX workshop on hot topics in cloud computing (HotCloud 16).
  17. The Early Microbenchmark Catches the Bug – Studying Performance Issues Using Micro- and Application Benchmarks. In Proceedings of the 16th IEEE/ACM International Conference on Utility and Cloud Computing. 1–10.
  18. PerfCI: A toolchain for automated performance testing during continuous integration of python projects. In Proceedings of the 35th IEEE/ACM International Conference on Automated Software Engineering (ASE ’20). 1344–1348.
  19. Tomas Kalibera and Richard Jones. 2020. Quantifying performance changes with effect size confidence intervals. arXiv:2007.10899 [stat.ME] (2020).
  20. Philipp Leitner and Jürgen Cito. 2016. Patterns in the chaos—a study of performance variation and predictability in public iaas clouds. ACM Transactions on Internet Technology 16, 3 (2016), 1–23.
  21. Serverless computing: state-of-the-art, challenges and opportunities. IEEE Transactions on Services Computing 16, 2 (2022), 1522–1539.
  22. Kristian Linnet. 2000. Nonparametric estimation of reference intervals by simple and bootstrap-based procedures. Clinical chemistry 46, 6 (2000), 867–869.
  23. On the variety of methods for calculating confidence intervals by bootstrapping. Journal of Animal Ecology 84, 4 (2015), 892–897.
  24. F. R. S. Rev. Samuel Horsley. 1772. The Sieve of Eratosthenes. Being an Account of His Method of Finding All the Prime Numbers. Philosophical Transactions 62 (1772), 327–347.
  25. Joel Scheuner. 2022. Performance Evaluation of Serverless Applications and Infrastructures. Ph. D. Dissertation. Chalmers University of Technology.
  26. CrossFit: Fine-grained Benchmarking of Serverless Application Performance across Cloud Providers. In Proceedings of the IEEE/ACM 15th International Conference on Utility and Cloud Computing (UCC ’22). 51–60.
  27. The Night Shift: Understanding Performance Variability of Cloud Serverless Platforms. In Proceedings of the 1st Workshop on SErverless Systems, Applications and MEthodologies (SESAME ’23). 27–33.
  28. Amazon Web Services. 2021. AWS re:Invent 2020: What’s new in serverless. https://www.youtube.com/watch?v=aW5EtKHTMuQ&t=339s. Last accessed: May 17, 2024.
  29. Srdjan Susnic. 2017. Machine Learning for Flappy Bird using Neural Network and Genetic Algorithm. https://github.com/ssusnic/Machine-Learning-Flappy-Bird. Last accessed: May 17, 2024.
  30. Iter8: Online experimentation in the cloud. In Proceedings of the ACM Symposium on Cloud Computing (SoCC). 289–304.
  31. Kieker: A framework for application performance monitoring and dynamic software analysis. In Proceedings of the 3rd ACM/SPEC International Conference on Performance Engineering (ICPE). 247–248.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com