Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Online Gradient-Based Caching Policy with Logarithmic Complexity and Regret Guarantees (2405.01263v2)

Published 2 May 2024 in cs.LG, cs.NI, and cs.OS

Abstract: Commonly used caching policies, such as LRU (Least Recently Used) or LFU (Least Frequently Used), exhibit optimal performance only under specific traffic patterns. Even advanced machine learning-based methods, which detect patterns in historical request data, struggle when future requests deviate from past trends. Recently, a new class of policies has emerged that are robust to varying traffic patterns. These algorithms address an online optimization problem, enabling continuous adaptation to the context. They offer theoretical guarantees on the regret metric, which measures the performance gap between the online policy and the optimal static cache allocation in hindsight. However, the high computational complexity of these solutions hinders their practical adoption. In this study, we introduce a new variant of the gradient-based online caching policy that achieves groundbreaking logarithmic computational complexity relative to catalog size, while also providing regret guarantees. This advancement allows us to test the policy on large-scale, real-world traces featuring millions of requests and items - a significant achievement, as such scales have been beyond the reach of existing policies with regret guarantees. To the best of our knowledge, our experimental results demonstrate for the first time that the regret guarantees of gradient-based caching policies offer substantial benefits in practical scenarios.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (36)
  1. Fundamental limits on the regret of online network-caching. Proceedings of the ACM on Measurement and Analysis of Computing Systems, 4(2):1–31, 2020.
  2. Selecting several samples from a single population. Australian Journal of Statistics, 14(3):231–239, 1972.
  3. {{\{{Cost-Aware}}\}}{{\{{WWW}}\}} proxy caching algorithms. In USENIX Symposium on Internet Technologies and Systems (USITS 97), 1997.
  4. Regret minimization for reserve prices in second-price auctions. IEEE Transactions on Information Theory, 61(1):549–564, 2014.
  5. Follow the leader if you can, hedge if you must. The Journal of Machine Learning Research, 15(1):1281–1316, 2014.
  6. A utility optimization approach to network cache design. IEEE/ACM Transactions on Networking, 27(3):1013–1027, 2019.
  7. It’s time to revisit LRU vs. FIFO. In 12th USENIX Workshop on Hot Topics in Storage and File Systems (HotStorage 20), 2020.
  8. Regret-optimal online caching for adversarial and stochastic arrivals. In EAI International Conference on Performance Evaluation Methodologies and Tools, pages 147–163. Springer, 2022.
  9. Herman Otto Hartley. Systematic sampling with unequal probability and without replacement. Journal of the American Statistical Association, 61(315):739–748, 1966.
  10. The computational power of optimization in online learning. In Proceedings of the forty-eighth annual ACM symposium on Theory of Computing, pages 128–141, 2016.
  11. Adaptive online prediction by following the perturbed leader. 2005.
  12. Characterization of storage workload traces from production Windows servers. In 2008 IEEE International Symposium on Workload Characterization, pages 119–128. IEEE, 2008.
  13. Understanding storage traffic characteristics on enterprise virtual desktop infrastructure. In Proceedings of the 10th ACM International Systems and Storage Conference, SYSTOR ’17, pages 13:1–13:11. ACM, 2017.
  14. An o (1) algorithm for implementing the lfu cache eviction scheme. arXiv preprint arXiv:2110.11602, 2021.
  15. ARC: A Self-Tuning, low overhead replacement cache. In 2nd USENIX Conference on File and Storage Technologies (FAST 03), 2003.
  16. Online caching with optimistic learning. In 2022 IFIP Networking Conference (IFIP Networking), pages 1–9. IEEE, 2022.
  17. Optimistic no-regret algorithms for discrete caching. Proceedings of the ACM on Measurement and Analysis of Computing Systems, 6(3):1–28, 2022.
  18. Online caching with optimal switching regret. In 2021 IEEE International Symposium on Information Theory (ISIT), pages 1546–1551. IEEE, 2021.
  19. Numerical optimization. Springer, 1999.
  20. Esbjörn Ohlsson. Coordination of samples using permanent random numbers. Business Survey Methods, pages 153–169, 1995.
  21. LeadCache: Regret-optimal caching in networks. Advances in Neural Information Processing Systems, 34:4435–4447, 2021.
  22. Online convex optimization for caching networks. IEEE/ACM Transactions on Networking, 28(2):625–638, 2020.
  23. Learning to cache with no regrets. In IEEE INFOCOM 2019-IEEE Conference on Computer Communications, pages 235–243. IEEE, 2019.
  24. Computational optimal transport: With applications to data science. Foundations and Trends® in Machine Learning, 11(5-6):355–607, 2019.
  25. Learning cache replacement with CACHEUS. In 19th USENIX Conference on File and Storage Technologies (FAST 21), pages 341–354, 2021.
  26. A quantitative study of recency and frequency based web cache replacement strategies. In Proceedings of the 11th communications and networking simulation symposium, pages 70–78, 2008.
  27. Shai Shalev-Shwartz et al. Online learning and online convex optimization. Foundations and Trends® in Machine Learning, 4(2):107–194, 2012.
  28. No-regret caching via online mirror descent. ACM Transactions on Modeling and Performance Evaluation of Computing Systems, 8(4):1–32, 2023.
  29. SNIA. SNIA iotta repository block I/O traces. http://iotta.snia.org/traces/block-io. Accessed: Jan. 2024.
  30. Learning relaxed belady for content distribution network caching. In 17th USENIX Symposium on Networked Systems Design and Implementation (NSDI 20), pages 529–544, 2020.
  31. Yves Tillé. Sampling and estimation from finite populations. John Wiley & Sons, 2020.
  32. Adaptive oracle-efficient online learning. Advances in Neural Information Processing Systems, 35:23398–23411, 2022.
  33. Projection onto the capped simplex. arXiv preprint arXiv:1503.01002, 2015.
  34. A large scale analysis of hundreds of in-memory cache clusters at twitter. In 14th USENIX Symposium on Operating Systems Design and Implementation (OSDI 20), pages 191–208, 2020.
  35. Z. Song et at. CDN traces. https://github.com/sunnyszy/lrb. Accessed: Jan. 2024.
  36. Z. Song et at. Twitter production cache traces. https://github.com/twitter/cache-trace. Accessed: Jan. 2024.

Summary

We haven't generated a summary for this paper yet.