Jisuanji kexue (Feb 2023)

Optimization and Deployment of Memory-Intensive Operations in Deep Learning Model on Edge

  • Peng XU, Jianxin ZHAO, Chi Harold LIU

DOI
https://doi.org/10.11896/jsjkx.20221100135
Journal volume & issue
Vol. 50, no. 2
pp. 3 – 12

Abstract

Read online

As a large amount of data is increasingly generated from edge devices,such as smart homes,mobile phones,and wearable devices,it becomes crucial for many applications to deploy machine learning modes across edge devices.The execution speed of the deployed model is a key element to ensure service quality.Considering a highly heterogeneous edge deployment scenario,deep learning compiling is a novel approach that aims to solve this problem.It defines models using certain DSLs and generates efficient code implementations on different hardware devices.However,there are still two aspects that are not yet thoroughly investigated yet.The first is the optimization of memory-intensive operations,and the second problem is the heterogeneity of the deployment target.To that end,in this work,we propose a system solution that optimizes memory-intensive operation,optimizes the subgraph distribution,and enables the compiling and deployment of DNN models on multiple targets.The evaluation results show the performance of our proposed system.

Keywords