Revolutionizing Deep Neural Networks on Mobile Platforms: MoDNN

Revolutionizing Deep Neural Networks on Mobile Platforms: MoDNN

At the recent DATE Design Automation Conference in Lausanne, Switzerland, a team of researchers from Duke University made a groundbreaking announcement. Led by Professor Yiran Chen, they unveiled MoDNN, a locally distributed mobile computing system designed to accelerate deep neural networks (DNN) on mobile platforms. This innovative solution has the potential to transform the way we interact with our smartphones and other mobile devices.

The Challenge of Deploying DNN on Mobile Devices

Deep neural networks have become ubiquitous in various applications, from image recognition to natural language processing. However, deploying DNN on resource-constrained devices like mobile platforms poses significant challenges. The computational requirements of DNN are substantial, and existing solutions often rely on client-server computing models or compression techniques that require specialized infrastructure.

MoDNN: A Locally Distributed Mobile Computing System

MoDNN proposes a novel approach to addressing this challenge. By dividing the DNN model into multiple mobile devices, MoDNN reduces the computational cost and memory usage of the calculation stage, thereby accelerating DNN computation. The researchers designed two models, BODP (Biased One-Dimensional Partition) and MSCC (Modified Spectral Co-Clustering), to minimize non-parallel data and reduce data transmission time.

Experimental Results

The experimental results demonstrate the effectiveness of MoDNN. When the number of nodes increases from 2 to 4, MoDNN accelerates DNN calculation by 2.17-4.28X. This significant improvement is achieved through the implementation of high parallelism and reduced data transfer time.

MoDNN System Framework

The MoDNN system framework consists of three main components:

  1. Local Distributed Network: A cluster formed by a plurality of operating nodes, with one node acting as the Group Owner (GO).
  2. Model Processor: The DNN model is processed on the working node.
  3. Intermediate Transfer Data and Identification Services: The services performed by DNN are identified and executed on the working node.

Advanced Partitioning Schemes

To balance the workload of each working node and minimize data transfer time, the authors propose several advanced partitioning schemes, including BODP, MSCC, and Fine-Grain CrossPartition (FGCP). These schemes are designed to optimize the execution time of the convolution layer and fully connected layer of the DNN.

Conclusion

MoDNN has the potential to revolutionize the way we interact with our smartphones and other mobile devices. By accelerating DNN computation on mobile platforms, MoDNN can improve the performance of mobile phone apps and enable more intelligent mobile devices. The researchers’ innovative approach to deploying DNN on resource-constrained devices has significant implications for the future of mobile computing.

Figure 1: MoDNN System Framework

The MoDNN system framework consists of three main components: Local Distributed Network, Model Processor, and Intermediate Transfer Data and Identification Services.

Technical Details

For more technical details on system implementation and experimental setup, please refer to the original paper.

Epilogue

In this paper, the authors proposed MoDNN, a locally distributed mobile computing system designed to accelerate DNN computation on mobile platforms. By dividing the DNN model into multiple mobile devices, MoDNN reduces the computational cost and memory usage of the calculation stage, thereby accelerating DNN computation. The experimental results demonstrate the effectiveness of MoDNN, and the authors propose several advanced partitioning schemes to optimize the execution time of the convolution layer and fully connected layer of the DNN.