The content discusses the TLMAC framework for processing quantised neural networks on FPGAs. It introduces the concept of Table Lookup Multiply-Accumulate (TLMAC) to compile and optimize quantised neural networks for scalable lookup-based processing. The framework clusters unique groups of weights to enable highly parallel computation, reducing LUT utilization and routing congestion. TLMAC significantly improves scalability compared to previous methods, allowing implementation of ImageNet-scale models on commercially available FPGAs.
To Another Language
from source content
arxiv.org
Thông tin chi tiết chính được chắt lọc từ
by Daniel Gerli... lúc arxiv.org 03-19-2024
https://arxiv.org/pdf/2403.11414.pdfYêu cầu sâu hơn