Notes On Quantized Convolutional Neural Networks

Published: 07 Jan 2016 Category: deep_learning

Existing works:

(1) Speed-up convolutional layers: Low-rank approximation and Tensor decomposition

(2) Reduce storage consumption in fully-connected layers: Parameter compression

Drawback: hard to achieve significant acceleration and compression simultaneously for the whole network.