Block-Based Compression and Corresponding Hardware Circuits for Sparse Activations
In a CNN (convolutional neural network) accelerator, to reduce memory traffic and power consumption, there is a need to exploit the sparsity of activation values. Therefore, some research efforts have been paid to skip ineffectual computations (i.e., multiplications by zero). Different from previous...
Saved in:
Main Authors: | Yui-Kai Weng, Shih-Hsu Huang, Hsu-Yu Kao |
---|---|
Format: | article |
Language: | EN |
Published: |
MDPI AG
2021
|
Subjects: | |
Online Access: | https://doaj.org/article/5cf53a5820af40cca4e6ee6645d2bf4d |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Learned Image Compression With Separate Hyperprior Decoders
by: Zhao Zan, et al.
Published: (2021) -
Design of universal convolutional layer IP core based on FPGA
by: Guochen AN, et al.
Published: (2021) -
Evaluation of Deep Neural Network Compression Methods for Edge Devices Using Weighted Score-Based Ranking Scheme
by: Olutosin Ajibola Ademola, et al.
Published: (2021) -
Sparse and dense matrix multiplication hardware for heterogeneous multi-precision neural networks
by: Jose Nunez-Yanez, et al.
Published: (2021) -
Power Efficient Tiny Yolo CNN Using Reduced Hardware Resources Based on Booth Multiplier and WALLACE Tree Adders
by: Fasih Ud Din Farrukh, et al.
Published: (2020)