The paper introduces HR-Cache, a learning-based caching framework that aims to optimize content delivery performance at the network edge. The key insights are:
HR-Cache is grounded in the principles of Hazard Rate Ordering (HRO), a rule that computes an upper bound on cache performance. It employs a lightweight machine learning model to learn from caching decisions made based on HRO, and subsequently predicts the "cache-friendliness" of incoming requests.
To accurately determine the hazard rate function for object inter-request times, HR-Cache uses a Kernel Hazard Estimator, which estimates the hazard function directly from the data without making simplifying assumptions about the nature of the request distribution.
HR-Cache divides its framework into two main components: 1) Calculating caching decisions for a window of past requests based on the HRO rule, and 2) Training an ML model to map object features to these HRO-based cache decisions.
During operation, HR-Cache preferentially evicts items in the cache that were previously identified as "cache-averse" based on the trained ML model's predictions.
Extensive experiments using three real-world traces and one synthetic trace demonstrate that HR-Cache consistently achieves 2.2–14.6% greater WAN traffic savings compared to the LRU replacement policy. It also outperforms not only heuristic caching strategies but also the state-of-the-art learning-based algorithm, LRB.
HR-Cache's design optimizations, such as batch-mode inference, reduce the prediction overhead by a factor of 19.2x compared to LRB, enhancing its practical efficiency.
Sang ngôn ngữ khác
từ nội dung nguồn
arxiv.org
Thông tin chi tiết chính được chắt lọc từ
by Hoda Torabi,... lúc arxiv.org 04-05-2024
https://arxiv.org/pdf/2402.02795.pdfYêu cầu sâu hơn