Skip to main content
eScholarship
Open Access Publications from the University of California

UC San Diego

UC San Diego Electronic Theses and Dissertations bannerUC San Diego

Resource Efficient and Error Resilient Neural Networks

Abstract

The entangled guardbands in terms of timing specification and energy budget ensure a system against faults, but the guardbands, meanwhile, impede the advance of a higher throughput and energy efficiency. To combat the over-designed guardbands in a system carrying out deep learning inference, we dive into the algorithmic demands and understand that the resource deficiency and hardware variation are the major reasons of the need of conservative guardbands. In modern convolutional neural networks (CNNs), the number of arithmetic operations for the inference could exceed tens of billions, which requires a sophisticated buffering mechanism to balance between resource utilization and throughput. In this case, the over-designed guardbands can seriously hinder system performance. On the other hand, timing errors can be incurred by the hardware variations including momentary voltage droops resulted from simultaneous switching noises, a gradually decreasing voltage level due to a limited battery, and the slow electron mobility incurred by the system power dissipation into heat. The timing errors propagating in a network can be a snowball in the beginning but ends up with a catastrophe in terms of a significant accuracy degradation.

Knowing the need of guardbands originates from resource deficiency and timing errors, this dissertation focuses on cross-layer solutions to the problems of the high algorithmic demands incurred by deep learning methods and error vulnerability due to hardware variations. We begin with reviewing the methods and technologies proposed in the literature including weight encoding, filter decomposition, network pruning, efficient structure design, and precision quantizing. In the implementation of an FPGA accelerator for extreme-case quantization, binarized neural networks (BNN), we have realized more possible optimizations can be applied. Then, we extend BNN on the algorithmic layer with the binarized separable filters and proposed BCNNw/SF. Although the quantization and approximation benefit hardware efficiency to a certain extent, the optimal reduction or compression rate is still limited by the core of the conventional deep learning methods -- convolution. We, thus, introduce the local binary pattern (LBP) to deep learning because of LBP's low complexity yet high effectiveness. We name the new algorithm LBPNet, in which the feature maps are created with a similar fashion of the traditional LBP using comparisons. Our LBPNet can be trained with the forward-backward propagation algorithm to extract useful features for image classification. LBPNet accelerators have been implemented and optimized to verify their classification performance, processing throughput, and energy efficiency. We also demonstrate the error immunity of LBPNet to be the strongest compared with the subject MLP, CNN, and BCNN models since the classification accuracy of the LBPNet is decreased by only 10% and all the other models lose the classification ability when the timing error rate exceeds 0.01.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View