Review On Machine Learning For Resource Usage Cost Optimization in Cloud Computing
Review On Machine Learning For Resource Usage Cost Optimization in Cloud Computing
https://doi.org/10.22214/ijraset.2023.51489
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue V May 2023- Available at www.ijraset.com
Abstract: Small and medium-sized enterprises are increasingly adopting cloud computing, and optimizing the cost of cloud
resources has become a crucial concern for them. Although several methods have been proposed to optimize cloud computing
resources, these methods mainly focus on a single factor, such as compute power, which may not yield satisfactory results in
real-world cloud workloads that are multi-factor, dynamic, and irregular.
This paper proposes a new approach that utilizes anomaly detection, machine learning, and particle swarm optimization to
achieve a cost-optimal cloud resource configuration.
The proposed solution works in a closed loop and does not require external supervision or initialization, learns about the
system's usage patterns, and filters out anomalous situations on the fly.
Additionally, the solution can adapt to changes in both system load and the cloud provider’s pricing plan. The proposed solution
was tested on Microsoft's Azure cloud environment using data collected from a real-life system, and the results show that it
achieved an 85% cost reduction over a ten-month period..
Index Terms: cloud resource usage prediction, anomaly detection, machine learning, particle swarm optimization, resource cost
optimization. .
I. INTRODUCTION
Cloud computing providers like Amazon Web Services (operated by Amazon), Azure (operated by Microsoft), and Google Cloud
Platform (operated by Google) are popular locations for computer systems.
These clouds offer storage, network, and computing resources to users who need them. Different cloud usage models, such as
Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service (SaaS), reduce management effort and
downtime risk while providing high scalability possibilities compared to on-premise solutions.
Scalability allows for the addition of new instances of services (PaaS), virtual machines (IaaS), or databases (which are partially
SaaS and partially PaaS) as needed. However, it can be challenging to predict load beforehand, making it difficult to meet
accessibility and responsiveness requirements.
Therefore, the system must be scaled up with a margin for unforeseen load spikes and long-term load changes, resulting in
considerable power and storage overprovisioning and unnecessary spending To reduce costs and protect the environment, it is
crucial to optimize cloud resource usage by predicting demand for different resources, such as CPU, memory, storage, and
input/output operations per second (IOPS), and adjusting cloud components accordingly. Our proposed solution automates the
process of scaling system components while taking into account the predicted usage level, including virtual machines, application
services, and databases.
We use machine learning interpolation combined with anomaly detection to predict demand and optimize cloud components that
meet the demand and are financially optimal. To achieve the optimal configuration, we use a particle swarm optimization (PSO)
algorithm tailored to solving discrete problems.
The traditional approach to cloud resource optimization either focuses on a single resource, such as CPU, and scaling parameter,
like the number of machines, or creates resource utilization models that ignore potential unexpected changes. Our proposed solution
takes a more comprehensive approach that considers all resources, predicts demand, and adjusts cloud components accordingly,
leading to significant cost reductions and environmental protection..
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 468
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue V May 2023- Available at www.ijraset.com
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 469
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue V May 2023- Available at www.ijraset.com
network traffic, and storage usage. You would then fit an autoregressive model with polynomial coefficient to this data, using
techniques such as least squares regression or maximum likelihood estimation. Once you have fitted the model, you can use it
to make predictions about future resource usage. For example, you might use the model to predict how much CPU capacity will
be needed in the next hour, based on past CPU utilization and other factors that affect demand, such as time of day or day of the
week. It's important to note that while autoregressive models with polynomial coefficients can be effective at forecasting time
series data, they may not be suitable for all types of demand forecasting problems. Other models, such as neural networks or
decision trees, may be more appropriate depending on the specific characteristics of the data and the problem you are trying to
solve.
4) The methodology for this paper involves proposing and evaluating three algorithms for cost optimization in cloud data centers.
The first algorithm is an optimal offline algorithm that leverages dynamic and linear programming techniques to minimize cost
under the assumption of available exact knowledge of workload on objects. The second and third algorithms are online
algorithms that dynamically select storage classes across CSPs while making trade-offs between residential and migration costs.
The methodology for this paper involves proposing and evaluating three algorithms for cost optimization in cloud data centers.
The first algorithm is an optimal offline algorithm that leverages dynamic and linear programming techniques to minimize cost
under the assumption of available exact knowledge of workload on objects. The second and third algorithms are online
algorithms that dynamically select storage classes across CSPs while making trade-offs between residential and migration costs.
A deterministic online algorithm is an algorithm that makes decisions without knowledge of future input. It operates in an
online setting where input arrives in a sequential order and decisions must be made immediately based on the available
information at the time. In a deterministic online algorithm, the decisions made are based solely on the input received up to that
point in time, without any consideration of future input. This is in contrast to a stochastic online algorithm, which may use
probabilistic methods to make decisions based on uncertain future input.Deterministic online algorithms are often used in real-
time systems, where decisions must be made quickly and efficiently based on real-time data. A randomized online algorithm is
an algorithm that makes decisions in an online setting using randomization. Like deterministic online algorithms, randomized
online algorithms operate in a sequential, online setting, where input arrives in a sequential order and decisions must be made
immediately based on the available information at the time
5) The methodology for this paper involves the use of a deep neural network as a function approximator for the Q-function in the
reinforcement learning algorithm. The authors use the experience replay technique to store and sample experiences for training
the neural network. They also use an epsilon-greedy policy for exploration and exploitation during training. The sensitivity
analysis involves varying different parameters such as the number of VMs, the workload intensity, and the learning rate to
evaluate their impact on the performance of the DRL approach. An intelligent resource management architecture, mainly
contains two components: a intelligent resource manager, which is composed of controller, monitor, and allocator and an IT
resource, which consists of extensive resource pools.12 Clients first communicate with the controller to submit application
requests with various demands. Based on application demands and current resource utilization information, the controller
implements the algorithm chosen from its resource schedule algorithm pool to meet application demands, while respecting
system resource constraint. The resource schedule algorithm pool, which plays an important role in intelligent resource
management architecture, includes different kinds of algorithms, such as offline and online algorithms and algorithms
combining both online and offline parts. The monitor is responsible for gathering information of system resource utilization and
application quality of service (QoS) to update the controller periodically, and the allocator is in charge of mapping applications
to resource pools according to the configuration negotiated by the controller. The controller is the key part of a resource
management architecture, as it not only figures out the (near-) optimal configuration policy but also coordinates with the
monitor and allocator to allocate resources intelligently. The heart of the controller is a resource schedule algorithm pool, which
contains plenty of control algorithms. The DRL algorithm presented in this paper is an online algorithm, which connects
reinforcement learning with deep learning to generate the (near-) optimal resource configuration in limited iterations directly
from raw application demands, especially for high dimensional demands.. The deep neural network is pretrained through the
stacked autoencoder (SA), followed by using reinforcement learning experiences for optimization.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 470
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue V May 2023- Available at www.ijraset.com
1. Tajwar 2018 The paper proposes an ensemble-based workload predictor ● Computational Overhead
Mehmood, using stacking mechanism to predict cloud computing ● Limited data set
Dr.Seemab resource utilization. The dataset used is the Google cluster ● No real world application
Latif ,Dr. usage trace data. The proposed model aims to improve
Sheheryaar resource utilization. Stacking is used to prune heterogeneous
Malik learning algorithms and reduce error of classifiers. KNN and
DT are base learners, and DT is the Meta learner.
2. Nirmal 2020 The paper propose a system to predict cloud computing ● Scalability
Kumawat, resources to serve input asset processing request . The ● Resource utilization
Nikhil Handa, method And system are designed in such a way to maximize ● Not Optimal long term
Avinash resource utilization, minimize cost spent and minimize strategy
Kharbanda processing time by such computational resources. The
method includes training of supervised learning based
predictive model with historic data which includes asset
processing requests, asset properties..
3. Quan Ding, 2018 The paper proposes using an autoregressive model with ● Lack of real-world testing
Bo Tang, polynomial coefficient to forecast the demand for cloud ● Limited Scope
Prakash computing resources. This model assumes a linear ● Lack of Comparison
Manden, Jin relationship between past and future values and a polynomial
Ren function to describe the relationship. Historical data on
resource usage would be collected, and the model would be
fitted to the data. The fitted model can then be used to predict
future resource usage.
4. Yaser 2019 The methodology for this paper involves proposing and ● Limited Scope
Mansouri, evaluating three algorithms for cost optimization in cloud ● Lack of real world
Adel Nadjaran data centers. The first algorithm is an optimal offline validation
Toosi , algorithm that leverages dynamic and linear programming ● Complexity
Rajkumar techniques to minimize cost under the assumption of
Buyya available exact knowledge of workload on objects. The
second and third algorithms are online algorithms that
dynamically select storage classes across CSPs while making
trade-offs between residential and migration costs.
5. Yu Zhang, 2018 This paper proposes a methodology for intelligent cloud ● Large amount of
Jianguo Yao, resource management using deep reinforcement learning. Computational Resources
Haibing Guan The authors formulate the problem as a Markov decision ● Complexity
process and develop a deep Q-network algorithm to learn an ● Large amount of training
optimal resource allocation policy. They evaluate the Data
performance using experiments on a cloud simulation
platform.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 471
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue V May 2023- Available at www.ijraset.com
V. ACKNOWLEDGEMENT
One's success cannot be solely attributed to their individual efforts as it is also influenced by the guidance, encouragement, and
cooperation of mentors, seniors, and companions. We express our gratitude to Dr. Praveen Kumar K V, a Professor in the
Computer Science and Engineering Department at Sapthagiri College of Engineering, and Dr. Kamalakshi Naganna, the Head of
the Computer Science and Engineering Department at Sapthagiri College of Engineering, for their unwavering backing, direction,
and aid during our project. Additionally, we extend our appreciation to our parents and friends for providing us with emotional
support throughout the journey.
REFERENCES
[1] T. Mehmood, S. Latif and S. Malik, "Prediction Of Cloud Computing Resource Utilization," 2018 15th International Conference on Smart Cities: Improving
Quality of Life Using ICT & IoT (HONET-ICT), Islamabad, Pakistan, 2018, pp. 38-42, doi: 10.1109/HONET.2018.8551339..
[2] N. Kumawat, N. Handa and A. Kharbanda, "Cloud Computing Resources Utilization and Cost Optimization for Processing Cloud Assets," 2020 IEEE
International Conference on Smart Cloud (SmartCloud), Washington, DC, USA, 2020, pp. 41-48, doi: 10.1109/SmartCloud49737.2020.00017.
[3] Q. Ding, B. Tang, P. Manden and J. Ren, "A learning-based cost management system for cloud computing," 2018 IEEE 8th Annual Computing and
Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 2018, pp. 362-367, doi: 10.1109/CCWC.2018.8301738.
[4] Y. Mansouri, A. N. Toosi and R. Buyya, "Cost Optimization for Dynamic Replication and Migration of Data in Cloud Data Centers," in IEEE Transactions on
Cloud Computing, vol. 7, no. 3, pp. 705-718, 1 July-Sept. 2019, doi: 10.1109/TCC.2017.2659728.
[5] Y. Zhang, J. Yao and H. Guan, "Intelligent Cloud Resource Management with Deep Reinforcement Learning," in IEEE Cloud Computing, vol. 4, no. 6, pp. 60-
69, November/December 2017, doi: 10.1109/MCC.2018.1081063.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 472