Core Concepts
FSD-Inference is a groundbreaking system for distributed ML inference, leveraging serverless computing and innovative communication channels to achieve scalability and cost-effectiveness.
Abstract
This content discusses the challenges of serverless computing for data-intensive applications and machine learning workloads. It introduces FSD-Inference as a fully serverless solution for distributed ML inference, detailing its design, communication schemes, optimizations, and cost models. The content also explores related work in the field of serverless computing.
Directory:
Abstract
Serverless computing benefits but limitations for data-intensive applications.
Introduction
Challenges of ML inference in cloud platforms.
Designing a Cloud-Based Serverless ML Solution
Key building blocks: compute engine, distributed processing, IPC patterns.
FSI Algorithm
Description of Fully Serverless Inference algorithms with communication channels.
FSD-Inference Cost Model
Breakdown of cost models for different communication channels.
FSD-Inference Optimizations
Strategies to reduce costs and improve performance.
Serverless Inference Design Recommendations
Recommendations for designing fully serverless ML inference systems.
Related Work
Overview of related research on serverless computing.
Stats
"FSD-Inference is significantly more cost-effective and scalable."
"Our solution achieves low latency and high throughput."
"The total number of Lambda instances is denoted by P."