The paper introduces HR-Cache, a learning-based caching framework that aims to optimize content delivery performance at the network edge. The key insights are:
HR-Cache is grounded in the principles of Hazard Rate Ordering (HRO), a rule that computes an upper bound on cache performance. It employs a lightweight machine learning model to learn from caching decisions made based on HRO, and subsequently predicts the "cache-friendliness" of incoming requests.
To accurately determine the hazard rate function for object inter-request times, HR-Cache uses a Kernel Hazard Estimator, which estimates the hazard function directly from the data without making simplifying assumptions about the nature of the request distribution.
HR-Cache divides its framework into two main components: 1) Calculating caching decisions for a window of past requests based on the HRO rule, and 2) Training an ML model to map object features to these HRO-based cache decisions.
During operation, HR-Cache preferentially evicts items in the cache that were previously identified as "cache-averse" based on the trained ML model's predictions.
Extensive experiments using three real-world traces and one synthetic trace demonstrate that HR-Cache consistently achieves 2.2–14.6% greater WAN traffic savings compared to the LRU replacement policy. It also outperforms not only heuristic caching strategies but also the state-of-the-art learning-based algorithm, LRB.
HR-Cache's design optimizations, such as batch-mode inference, reduce the prediction overhead by a factor of 19.2x compared to LRB, enhancing its practical efficiency.
翻譯成其他語言
從原文內容
arxiv.org
深入探究