event

PhD Proposal by Ramyad Hadidi

Primary tabs

Title: Deploying Deep Neural Networks in Edge with Distribution

------------

 

Ramyad Hadidi

Ph.D. Student

School of Computer Science

College of Computing

Georgia Institute of Technology

 

Date: Thursday, December 19, 2019

Time: 11:30 AM - 1:30 PM (EST)

Location: Klaus 2100

 

Committee:

------------

Dr. Hyesoon Kim (Advisor, School of Computer Science, Georgia Institute of Technology)

Dr. Saibal Mukhopadhyay (School of Electrical and Computer Engineering, Georgia Institute of Technology)

Dr. Tushar Krishna (School of Electrical and Computer Engineering, Georgia Institute of Technology)

Dr. Alexey Tumanov (School of Computer Science, Georgia Institute of Technology)

 

 

Abstract:

------------

The widespread applicability of deep neural networks (DNNs) has led edge computing to emerge as a trend to extend our capabilities to several domains such as robotics, autonomous technologies, and Internet-of-things devices. Because of the tight resource constraints of such individual edge devices, computing accurate predictions while providing a fast execution is a key challenge. Moreover, modern DNNs increasingly demand more computation power than their predecessors. As a result, the current approach is to rely on compute resources in the cloud by offloading the inference computations of DNNs. This approach not only does raise privacy concerns but also relies on network infrastructure and data centers that are not scalable and do not guarantee fast execution.

 

Our key insight is that edge devices can break their individual resource constraints by distributing the computation of DNNs on collaborating peer edge devices. In our approach, edge devices cooperate to conduct single-batch inferences in real-time while exploiting several model-parallelism methods. Nonetheless, since communication is costly and current DNN models capture a single-chain of dependency pattern, distributing and parallelizing the computations of current DNNs may not be an effective solution for edge domains. Therefore, to efficiently benefit from computing resources with low communication overhead, we propose new handcrafted edge-tailored models that consist of several independent and narrow DNNs. Additionally, we explore an automated neural architecture search methodology and propose ParallelNets, custom DNN architectures with low communication overheads and high parallelization opportunities.

Status

  • Workflow Status:Published
  • Created By:Tatianna Richardson
  • Created:12/10/2019
  • Modified By:Tatianna Richardson
  • Modified:12/11/2019

Categories

Keywords