Show simple item record

Customized Systems of Deep Neural Networks for Energy Efficiency

dc.contributor.authorZamirai, Babak
dc.date.accessioned2020-10-04T23:38:27Z
dc.date.available2022-09-01
dc.date.available2020-10-04T23:38:27Z
dc.date.issued2020
dc.identifier.urihttps://hdl.handle.net/2027.42/163278
dc.description.abstractThe size and complexity growth of deep neural networks (DNNs), which is driven by the push for higher accuracies and wider ranges of functionality, is outpacing the growth of hardware for mobile systems. Thus, edge devices must collect and transmit sensory data for analysis on a remote server. However, offloading the data to be processed on the backend server can be problematic due to the high cost/latency of wireless communication of the raw data as well as potential privacy/security concerns. Hence, there has been significant research interest to increase performance and energy efficiency of deep learning computation on both mobile devices and servers. These techniques can be divided into two major categories: input-invariant and input-variant. Input-invariant methods leverage hardware and software techniques, such as pruning, to accelerate a single DNN for the entire dataset. On the other hand, input-variant approaches focus on the fact that most of the inputs do not require the entire computational power of the model to produce an accurate final output. Consequently, they employ a combination of simple and complex models to dynamically adjust the complexity of the DNN to input difficulty. Both DNN pruning and intelligent combination of DNNs require machine learning expertise and manual design, which makes them hard to implement and deploy. This thesis proposes techniques to improve performance and applicability of both input-invariant and input-variant methods. First, it introduces a new category of input-variant solutions to maximize energy efficiency and minimize latency of DNNs by customizing systems of DNNs based on input variations. Instead of conventional DNN ensembles, it proposes a data heterogeneous multi-NN system to divide the data space into subsets with one specialized learner for each subset. In addition, an intelligent hybrid server-edge deep learning system is introduced to dynamically distribute DNN computation between the cloud and edge device based on the input data and environmental conditions. Furthermore, it suggests an input-driven synergistic deep learning system, which dynamically distributes DNN computation between a more accurate big and a less accurate little DNN. At the end, it introduces a noniterative double-shot pruning method, which takes advantage of both architectural features and weight values to improve the simplicity and applicability of pruning. Compared to the conventional approaches, on average, an energy consumption reduction up to 91% and a performance improvement up to 12.3x are achieved, while maintaining the original accuracy.
dc.language.isoen_US
dc.subjectEfficient Deep Neural Network
dc.subjectEfficient Deep Learning
dc.titleCustomized Systems of Deep Neural Networks for Energy Efficiency
dc.typeThesis
dc.description.thesisdegreenamePhDen_US
dc.description.thesisdegreedisciplineComputer Science & Engineering
dc.description.thesisdegreegrantorUniversity of Michigan, Horace H. Rackham School of Graduate Studies
dc.contributor.committeememberMahlke, Scott
dc.contributor.committeememberKim, Hun Seok
dc.contributor.committeememberDreslinski Jr, Ronald
dc.contributor.committeememberTang, Lingjia
dc.subject.hlbsecondlevelComputer Science
dc.subject.hlbtoplevelEngineering
dc.description.bitstreamurlhttp://deepblue.lib.umich.edu/bitstream/2027.42/163278/1/zamirai_1.pdfen_US
dc.identifier.orcid0000-0001-8379-0834
dc.identifier.name-orcidZamirai, Babak; 0000-0001-8379-0834en_US
dc.owningcollnameDissertations and Theses (Ph.D. and Master's)


Files in this item

Show simple item record

Remediation of Harmful Language

The University of Michigan Library aims to describe its collections in a way that respects the people and communities who create, use, and are represented in them. We encourage you to Contact Us anonymously if you encounter harmful or problematic language in catalog records or finding aids. More information about our policies and practices is available at Remediation of Harmful Language.

Accessibility

If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.