Memory Access Optimization for On-Chip Transfer Learning
Abstract:
Training of Deep Neural Network (DNN) at the edge
faces the challenge of high energy consumption due to the requirements of a large number of
memory accesses for gradient calculations. Therefore, it is necessary to minimize data fetches
to perform training of a DNN model on the edge. In this paper, a novel technique has been proposed
to reduce the memory access for the training of fully connected layers in transfer learning. By analyzing
the memory access patterns in the backpropagation phase in fully connected layers, the memory access can be
optimized. We introduce a new method to update the weights by introducing the delta term for every node of
output and fully connected layer. Delta term aims to reduce memory access for the parameters which are required
to access repeatedly during the training process of fully connected layers. The proposed technique shows 0.13x-13.93x
energy savings for the training of fully connected layers for famous DNN architectures on multiple processor architectures.
The proposed technique can be used to perform transfer learning on-chip to reduce energy consumption as well as memory access.
Details:
Following equation is used for weight update in FC layers
In the proposed method, we use following equation to optimize the weight update process in fully connected layers.
Result:
Made
by ¼Ú®ü¬À