Optimizing Resource Allocation in Cloud for Large-Scale Deep Learning Models in Natural Language Processing

Gauri Dhopavkar,Rashmi R. Welekar,Piyush K. Ingole, Chandu Vaidya, Shalini Vaibhav Wankhade, Bharati P. Vasgi

JOURNAL OF ELECTRICAL SYSTEMS(2023)

引用 0|浏览2
暂无评分
摘要
The need for big deep learning models in Natural Language Processing (NLP) keeps rising, it's important to find the best way to divide up cloud resources so that they can be used efficiently and at high speeds. This solves the problems that come with setting up and handling large NLP models by suggesting a complete strategy for making the best use of cloud-based platforms' resources. Combining model parallelism, data parallelism, and dynamic scaling methods, the suggested approach spreads the computing load across multiple cloud instances in better way. The framework constantly changes how resources are allocated to handle changes in workload by taking into account the specifics of NLP tasks, such as the need for different model designs and data processing needs. To improve scale and cut down on inference delay, a new auto-scaling method is introduced that lets computing resources be changed automatically based on demand in real time. The framework uses machine learning-based prediction models to figure out what resources will be needed in the future. This lets you make proactive decisions about scaling and keeps you from underusing or overprovisioning resources. It also solves the problem of communication overhead in distributed environments by improving data exchange protocols and using advanced inter-process communication techniques. The results of the experiments show that the proposed framework works well at improving both cost-effectiveness and prediction performance for large-scale NLP models by making the best use of resources. The framework is flexible enough to work with a wide range of natural language processing (NLP) tasks. It makes a useful addition to the efficient use of deep learning models in cloud settings.
更多
查看译文
关键词
Resource Allocation,NLP,Deep Learning,Optimization
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要