Allocating Resource Capacities for an Offload-enabled Mobile Edge Cloud System

2022 IEEE Eighth International Conference on Big Data Computing Service and Applications (BigDataService)(2022)

引用 0|浏览4
暂无评分
摘要
AI applications have become increasingly popular and been widely deployed in various scenarios. In an AI application, an AI model is typically first trained, and then the model inference service is deployed in systems to perform model inference tasks based on the input data. In this paper, we systematically model the performance of running such AI tasks on Mobile Edge Cloud (MEC) systems. In particular, the model inference services are deployed in mobile devices, the edge devices and the cloud server. Mobile devices collect the monitoring data and perform the model inference tasks. When the arrival rate of the incoming data becomes too big, the mobile devices offload a portion of model inference tasks to the edge devices by the way of uploading the incoming data to the edge devices. If an edge device is overwhelmed by the tasks, it can further offload to the cloud server. This paper aims to model the offloading behaviors in MEC and also the resource capacities required to meet the desired task performance.
更多
查看译文
关键词
mobile edge computing,AI applications,task performance,queuing theory
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要