Autor: |
Dhopavkar, Gauri, Welekar, Rashmi R., Ingole, Piyush K., Vaidya, Chandu, Wankhade, Shalini Vaibhav, Vasgi, Bharati P. |
Předmět: |
|
Zdroj: |
Journal of Electrical Systems; 2023, Vol. 19 Issue 3, p62-77, 16p |
Abstrakt: |
The need for big deep learning models in Natural Language Processing (NLP) keeps rising, it's important to find the best way to divide up cloud resources so that they can be used efficiently and at high speeds. This solves the problems that come with setting up and handling large NLP models by suggesting a complete strategy for making the best use of cloud-based platforms' resources. Combining model parallelism, data parallelism, and dynamic scaling methods, the suggested approach spreads the computing load across multiple cloud instances in better way. The framework constantly changes how resources are allocated to handle changes in workload by taking into account the specifics of NLP tasks, such as the need for different model designs and data processing needs. To improve scale and cut down on inference delay, a new auto-scaling method is introduced that lets computing resources be changed automatically based on demand in real time. The framework uses machine learning-based prediction models to figure out what resources will be needed in the future. This lets you make proactive decisions about scaling and keeps you from underusing or overprovisioning resources. It also solves the problem of communication overhead in distributed environments by improving data exchange protocols and using advanced inter-process communication techniques. The results of the experiments show that the proposed framework works well at improving both cost-effectiveness and prediction performance for large-scale NLP models by making the best use of resources. The framework is flexible enough to work with a wide range of natural language processing (NLP) tasks. It makes a useful addition to the efficient use of deep learning models in cloud settings. [ABSTRACT FROM AUTHOR] |
Databáze: |
Complementary Index |
Externí odkaz: |
|