StitchNet: Distributed On-Device Model Partitioning Over Edge Devices Under Volatile Wireless Links

Jiho Lee, Jeihee Cho, Hyungjune Lee

Research output: Contribution to journalArticlepeer-review

Abstract

Distributed deep learning architecture can achieve scalable learning and inference capability at resource-constrained edge devices. Although the parallelization-based approaches have actively been investigated in the edge computing context, they are not designed for the devices that are usually wireless and mobile, causing the substantial link and device failure issue. We propose a semi-distributed deep learning architecture, StitchNet, based on model parallelism for volatile wireless edge networks. Our algorithm first classifies a set of effective neurons with a substantial impact on their connected neurons across layers. Then, an opportunistic neuron assignment is employed to ensure the full forward and backward propagation paths by stitching the subsets of the model across the devices with path redundancy via neuron cloning for securing high resilience to network and device uncertainty. Simulation-based experiments demonstrate that StitchNet has achieved high inference quality on visual classification tasks even under the volatile lossy network environment, by making edge devices collaboratively find and stitch their in-ward and out-ward edge paths with a reasonable communication overhead.

Original languageEnglish
Pages (from-to)110616-110627
Number of pages12
JournalIEEE Access
Volume10
DOIs
StatePublished - 2022

Keywords

  • distributed learning
  • edge computing
  • model parallelism
  • On-device AI

Fingerprint

Dive into the research topics of 'StitchNet: Distributed On-Device Model Partitioning Over Edge Devices Under Volatile Wireless Links'. Together they form a unique fingerprint.

Cite this