CNN Accelerator Adapted to Quasi Structured Pruning and Dense Mode

Publish Year: 1404
نوع سند: مقاله ژورنالی
زبان: English
View: 78

This Paper With 15 Page And PDF Format Ready To Download

  • Certificate
  • من نویسنده این مقاله هستم

استخراج به نرم افزارهای پژوهشی:

لینک ثابت به این Paper:

شناسه ملی سند علمی:

JR_ITRC-17-3_003

تاریخ نمایه سازی: 30 شهریور 1404

Abstract:

In recent years, Convolutional Neural Networks (CNN) have been extensively used in machine learning algorithms related to images due to their exceptional accuracy. The multiplication-accumulation (MAC) in convolutional layers makes them computationally expensive, and these layers account for ۹۰% of the total computation. Several researchers have taken advantage of pruning the weights and activations to overcome high computation bandwidth. These techniques are divided into two categories: ۱) unstructured pruning of the weights can achieve heavy pruning, but in the process, it unbalances data access and computation processes. Consequently, compression coding for indexing non-zero data increases, which causes much more memory volume. ۲) Structured pruning by the specified pattern prunes the weights and regularizes both computations and memory access but does not support high pruning amounts compared to unstructured pruning. In this paper, we proposed Quasi Structured Pruning (QSP) that profits from the high pruning ratio of unstructured pruning. The load balancing property in structured pruning has also been included in the QSP scheme. Implementation results of our accelerator using VGG۱۶ on a Xilinx XC۷Z۱۰۰ indicate ۶۱۶.۹۴ GOP/s and ۱۴۳۷.۷ GOP/s at just ۷.۸ watts power consumption for dense and sparse mode, respectively. Experimental results show that the accelerator is ۱.۳۸×, ۱.۱×, ۲.۷۷×, ۲.۸۷×, ۱.۹۱×, and ۱.۱۸× better in terms of DSP efficiency than previous accelerators in dense mode. As well, our accelerator has achieved ۱.۹×, ۲.۹۲×, ۱.۶۷×, and ۱.۱۱× higher DSP efficiency besides ۴.۵۲×, ۵.۳۱×, ۱۰.۳۸×, and ۱.۱× better energy efficiency than other state-of-the-art sparse accelerators.

Keywords:

Authors

Amirhossein Sadough

Department of AI, Donders Center for Cognition, Radboud University Netherlands

Parviz Amiri

Department of Electrical Engineering Shahid Rajaee Teacher Training University Tehran, Iran

Mohammad Hossein Maghami

Department of Electrical Engineering Shahid Rajaee Teacher Training University Tehran, Iran