TabConv: Low-Computation CNN Inference via Table Lookups
arxiv(2024)
摘要
Convolutional Neural Networks (CNNs) have demonstrated remarkable ability
throughout the field of computer vision. However, CNN inference requires a
large number of arithmetic operations, making them expensive to deploy in
hardware. Current approaches alleviate this issue by developing
hardware-supported, algorithmic processes to simplify spatial convolution
functions. However, these methods still heavily rely on matrix multiplication,
leading to significant computational overhead. To bridge the gap between
hardware, algorithmic acceleration, and approximate matrix multiplication, we
propose TabConv, a novel, table-based approximation for convolution to
significantly reduce arithmetic operations during inference. Additionally, we
introduce a priority masking technique based on cosine similarity to select
layers for table-based approximation, thereby maintaining the model
performance. We evaluate our approach on popular CNNs: ResNet-18, ResNet-34,
and NetworkInNetwork (NIN). TabConv preserves over 93
performance while reducing arithmetic operations by 36.5
ResNet-18 on CIFAR-10, CIFAR-100, and MNIST, respectively, 35.6
ResNet-34 on CIFAR-10 and MNIST, and 98.9
low-computation inference.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要