Software Engineering

p-ISSN: 2162-934X    e-ISSN: 2162-8408

2024;  11(2): 11-18

doi:10.5923/j.se.20241102.01

Received: Sep. 19, 2024; Accepted: Oct. 3, 2024; Published: Oct. 14, 2024

 

Enhancing Data Analytics Using AI-Driven Approaches in Cloud Computing Environments

Ravi Kumar1, Neha Thakur2, Ahmad Saeed3, Chandra Jaiswal4

1Cloud Data & AI, Dollar General Corporation, Charlotte, NC, USA

2School of Computer Science and Engineering, Galgotias University, Greater Noida, India

3Stock Plan Services, Fidelity Investments. Durham, NC, USA

4Computational Data Science and Engineering, North Carolina Agricultural and Technical University, Greensboro, NC, USA

Correspondence to: Neha Thakur, School of Computer Science and Engineering, Galgotias University, Greater Noida, India.

Email:

Copyright © 2024 The Author(s). Published by Scientific & Academic Publishing.

This work is licensed under the Creative Commons Attribution International License (CC BY).
http://creativecommons.org/licenses/by/4.0/

Abstract

Data Science and Artificial Intelligence (AI) are revolutionizing the way organizations extract meaningful insights from vast datasets. Traditional data analysis methods face challenges in handling the growing data volume and complexity. AI-driven techniques, such as machine learning and deep learning, provide more efficient ways to process and analyze this data. Cloud computing, with its scalable and flexible infrastructure, complements AI by offering on-demand resources to manage large-scale data analytics. This paper examines the integration of AI into cloud computing to enhance data analysis. It explores how AI models can optimize cloud-based analytics, improving performance, scalability, and cost-efficiency. Through experiments and case studies, we compare traditional data analysis methods with AI-driven approaches, highlighting the advantages of real-time processing and automation in the cloud. The findings reveal that AI significantly boosts the efficiency of data processing in cloud environments, though challenges such as security and data privacy remain. We conclude with a discussion on the future potential of AI-powered cloud data analytics and its role in enabling faster, more accurate decision-making across various industries.

Keywords: Data Science, Artificial Intelligence, Machine Learning, Deep Learning, Cloud Computing

Cite this paper: Ravi Kumar, Neha Thakur, Ahmad Saeed, Chandra Jaiswal, Enhancing Data Analytics Using AI-Driven Approaches in Cloud Computing Environments, Software Engineering, Vol. 11 No. 2, 2024, pp. 11-18. doi: 10.5923/j.se.20241102.01.

1. Introduction

In the modern digital era, the exponential growth of data across industries has created an increasing demand for efficient data analysis techniques. Data Science and Artificial Intelligence have become pivotal in managing, processing, and extracting valuable insights from vast datasets. AI technologies, particularly machine learning and deep learning, have shown immense potential in transforming traditional data analysis by enabling predictive analytics, anomaly detection, and real-time decision-making.
However, as data continues to grow in both volume and complexity, traditional on-premises infrastructure often struggles to scale effectively. This is where cloud computing offers a significant advantage, providing scalable, elastic, and cost-effective infrastructure to handle large-scale data processing. The cloud's ability to deliver computing resources on demand makes it an ideal platform for hosting AI-driven data analytics solutions. By leveraging the combination of AI and cloud computing, organizations can harness the power of data more effectively while maintaining flexibility and reducing costs.
While AI and cloud computing are both transformative technologies, their integration in data analytics is still evolving. AI models require significant computational resources for training and inference, which often limits their deployment in traditional environments. Cloud computing offers a solution by providing scalable infrastructure, but integrating AI models with cloud services introduces new challenges, such as managing latency, cost-efficiency, and ensuring data security and privacy.
This paper seeks to fill these challenges by proposing an integrated approach that not only addresses the technical aspects of AI and cloud computing but also evaluates the practical implications of this integration, including performance, cost, and security. The significant contribution of this research work is:
• Present a novel framework for seamlessly integrating AI techniques, such as machine learning and deep learning, with cloud computing platforms (AWS) to enhance the efficiency and scalability of data analysis processes.
• Comprehensive experimental evaluation of the framework, assessing key metrics such as speed, scalability, model accuracy, and cost-efficiency. Our results offer valuable insights into the balance between computational power and cloud costs, allowing organizations to make informed decisions in optimizing their AI implementations.
This research not only presents an innovative architecture but also provides a pathway for future cloud-based AI solutions to handle large-scale data challenges while remaining efficient and scalable.
The remainder of this paper is structured as follows: Section 2 introduces the existing techniques of AI, Cloud Computing, and Data Analysis, highlighting their strengths and weaknesses. Section 3 outlines the proposed framework, detailing its design, components, and integration with cloud platforms. Section 4 discusses the experimental setup and results, offering a performance comparison between the proposed model and existing solutions. Finally, Section 5 presents the conclusion, summarizing key findings and suggesting directions for future work.

2. Literature Review

The convergence of AI and Cloud Computing in data analysis has gained significant attention in recent years, with researchers exploring how these technologies can optimize data processing and decision-making. This section provides a comprehensive review of existing studies, focusing on AI techniques for data analysis, cloud computing infrastructure for big data, and the challenges and opportunities of integrating AI with cloud platforms.

2.1. AI in Data Analytics

AI techniques, particularly machine learning (ML) and deep learning (DL), have revolutionized traditional data analysis by enabling predictive models, automation, and real-time decision-making. [1,2] highlights the power of machine learning in data-intensive applications, demonstrating its ability to process large datasets and discover patterns that would otherwise be difficult to uncover manually. Similarly, [3-5] describe deep learning’s role in improving the accuracy of data analysis, especially for unstructured data like images and text.
Despite these advancements, Dean et al. (2012) argue that the computational demands of AI models, especially deep learning, can create bottlenecks in resource-constrained environments. Training AI models often requires substantial computational power, which is typically beyond the capabilities of traditional on-premises data centers [6-8]. This gap in infrastructure presents an opportunity for cloud computing to address the scalability and computational limitations.

2.2. Cloud Computing for Scalable Data Processing

Cloud computing offers a solution to the scalability challenges of AI-driven data analytics by providing elastic, on-demand computing resources. [1,9,10] emphasize the key benefits of cloud computing, including scalability, flexibility, and cost efficiency, which make it ideal for handling large-scale data analytics workloads. Cloud platforms like Amazon Web Services (AWS), Microsoft Azure, and Google Cloud Platform (GCP) provide extensive infrastructure for storage and computation, which can be dynamically scaled to meet the needs of data analysis tasks.
The literature also shows that cloud services enable distributed data processing through models like MapReduce and Apache Spark, allowing organizations to process vast datasets in parallel. However, [11] note that while cloud computing is highly effective for large-scale computations, its potential in AI-driven data analysis has not been fully exploited due to challenges in optimizing resource allocation, minimizing latency, and ensuring data privacy.

2.3. Challenges in AI-Cloud Integration

Several studies have explored the integration of AI and cloud computing, though much of the research focuses on specific aspects rather than a holistic approach. [12,13] discuss the need for optimized AI deployment strategies in the cloud, emphasizing that efficient resource management is critical for achieving scalability without excessive costs. [8,14] provide a detailed analysis of cloud-based AI workflows, discussing how cloud services can accelerate model training and deployment but also highlighting concerns related to latency, especially when working with real-time data.
Security and data privacy are also major concerns in the integration of AI with cloud platforms [15,16]. [17] identify potential vulnerabilities when sensitive data is processed on cloud infrastructure, calling for stronger encryption techniques and data governance policies. Similarly, [4,18] explore privacy-preserving AI algorithms that can be integrated with cloud services to mitigate these risks.
Despite the growing body of research, there remains a lack of comprehensive studies focusing on optimizing the complete lifecycle of AI-driven data analytics in cloud environments. Most literature focuses on either enhancing AI models or improving cloud infrastructure independently [13,19-26]. Few studies offer practical frameworks that integrate both technologies to address issues of scalability, performance, and cost-efficiency. Furthermore, the challenges related to latency, security, and data privacy in cloud-based AI deployments are often discussed in isolation rather than as part of an integrated system.
This gap presents an opportunity for further research into creating more effective AI-cloud integration models, optimizing resource allocation, and addressing the trade-offs between computational efficiency, cost, and security. Our research seeks to contribute to this space by proposing a framework for AI-driven data analytics in cloud environments and evaluating its performance through real-world case studies.

3. Proposed Framework

The integration of Artificial Intelligence (AI) with Cloud Computing offers a transformative approach to data analysis by combining AI’s predictive and analytic capabilities with the cloud’s scalable and flexible infrastructure. While individual advancements in AI models and cloud services have significantly enhanced data processing, there is a growing need for a cohesive framework that optimizes AI-driven data analytics within cloud environments. Our proposed framework should not only streamline the deployment of AI models on cloud platforms but also address key challenges such as latency, resource allocation, cost-efficiency, and security.
We propose a comprehensive framework for integrating AI techniques with cloud computing platforms to enhance data analysis. The framework is designed to leverage cloud infrastructure’s elasticity to handle the computational demands of AI models, while also ensuring efficient data processing, resource optimization, and real-time insights. The proposed framework incorporates a layered architecture that ensures seamless communication between the AI models, cloud services, and data pipelines. It also addresses challenges such as ensuring data security, optimizing cloud costs, and minimizing latency in real-time analytics. The proposed framework is structured into five key layers:
• Data Ingestion Layer:
Function: The data ingestion layer is responsible for collecting and preparing data from multiple sources, such as IoT devices, databases, and streaming platforms. This layer uses cloud services like Amazon Kinesis, Azure Event Hubs, or Google Pub/Sub to collect real-time data and batch data from various sources.
Tools: Cloud-native tools such as AWS Glue or Azure Data Factory can be used for ETL (Extract, Transform, Load) processes to clean, transform, and load the data into a storage system, ensuring that data is ready for analysis.
• Data Storage Layer:
Function: This layer provides a scalable and cost-efficient storage system for large volumes of data. It uses cloud storage solutions like Amazon S3, Azure Blob Storage, or Google Cloud Storage, ensuring that structured, semi-structured, and unstructured data is stored securely and efficiently.
Tools: Storage can be coupled with data warehousing services like Amazon Redshift or Google BigQuery to provide query capabilities on large datasets. Data encryption and access control are implemented to ensure security and privacy compliance.
• AI Model Layer
Function: This is the core layer where AI-driven data analysis takes place. AI models (e.g., machine learning, deep learning) are designed, trained, and deployed in the cloud environment. Cloud platforms provide AI services such as AWS SageMaker, Azure Machine Learning, and Google AI Platform Vertex AI that offer pre-built and customizable AI models.
Tools: The layer supports model training using cloud compute services such as AWS EC2, Azure Virtual Machines, or Google Compute Engine. For AI-specific tasks, GPU/TPU instances can be provisioned to accelerate deep learning model training. Model deployment and inference services allow for scaling based on demand, ensuring that the models can handle large datasets efficiently.
• Processing and Orchestration Layer
Function: This layer orchestrates the data flow, ensuring that data is processed efficiently across different cloud services and AI models. It also manages the computational resources to balance workloads and optimize cost-efficiency. The orchestration services automatically scale resources based on the complexity of the AI tasks and incoming data.
Tools: Tools such as Kubernetes or AWS Lambda (for serverless computing) are used to manage and automate the deployment of AI models and data processing tasks. Apache Airflow or AWS Step Functions are used for managing and orchestrating data pipelines and workflows, ensuring that data is processed in real-time or near-real-time.
• Output and Visualization Layer
Function: Once the AI models have processed the data, this layer is responsible for delivering insights and visualizations to end-users in a comprehensible manner. It ensures that real-time dashboards, reports, and alerts are available for decision-makers.
Tools: Visualization and analytics tools like Tableau, Power BI, or Google Data Studio can be integrated to visualize the outcomes of the data analysis. Additionally, API services can be used to expose the results to external systems or applications for further use.

4. Experiments and Results

In this section, we present the experiments conducted to evaluate the performance of the proposed AI-driven data analytics framework in a cloud computing environment. The experiments were designed to assess the framework's ability to handle large-scale data processing, model training efficiency, scalability, cost optimization, and real-time data analysis. We utilized cloud platforms such as Amazon Web Services (AWS) and Google Cloud Platform (GCP) to implement and test the framework, focusing on performance metrics such as data processing speed, model accuracy, cost, and scalability.

4.1. Experiment Setup and Implementation Details

The entire document should be in Times New Roman. The font sizes to be used are specified in Table 1.
Table 1. Comparison of Proposed Model with Existing Techniques
     
The size of a lower-case “j” will give the point size by measuring the distance from the top of an ascender to the bottom of a descender.
1. Dataset and Data Ingestion
For our experiments, we used two publicly available datasets:
• Kaggle Credit Card Fraud Detection Dataset (284,807 transactions with 30 features): This dataset was chosen to evaluate how the framework handles supervised learning for binary classifications as shown in Figure 1 and Figure 2.
• IoT Sensor Data Stream from Intel Berkeley Research Lab (54 sensors with 2.3 million readings): This dataset was used to test the framework's real-time data ingestion and analysis capabilities.
Figure 1. Fraud detection dataset
Figure 2. Attributes of dataset
The data ingestion layer was implemented using AWS Kinesis for real-time stream data and AWS Glue for batch data processing. Both datasets were ingested into the cloud environment through these services, ensuring efficient ETL (Extract, Transform, Load) operations. The pre-processed data was then stored in Amazon S3 for scalable storage, allowing for both structured and unstructured data to be queried and analyzed.
2. AI Model Layer Implementation
For the credit card fraud detection dataset, we implemented a Random Forest and XGBoost classifier using AWS SageMaker. The models were trained on the dataset using a distributed computing environment powered by EC2 instances with GPU acceleration. The training process was automated using SageMaker’s hyperparameter tuning to optimize model performance.
For the IoT sensor dataset, a Long Short-Term Memory (LSTM) deep learning model was implemented using Google AI Platform to predict anomalies in sensor readings. We used Google Cloud TPU (Tensor Processing Unit) for accelerated training, given the large size of the dataset.
The model was designed to handle time-series data, learning from previous patterns to predict future sensor behaviors and detect anomalies. For data preprocessing and model evaluation, we leveraged Google BigQuery and Cloud Dataflow to efficiently handle and analyze the streaming data.
3. Processing and Orchestration Layer
We used AWS Lambda for serverless orchestration of data pipelines. This allowed for real-time processing of streaming data without provisioning dedicated servers. Apache Airflow was utilized for workflow automation, ensuring that the data was efficiently moved between storage, AI model training, and analysis.
Lambda functions automatically triggered the model inference when new data was ingested, enabling near real-time predictions as shown in Figure 3.
Figure 3. Lambda functions automatically triggered the model inference when new data was ingested
4. Output and Visualization Layer
The results of the AI models were visualized using Tableau for dashboard creation and AWS QuickSight for reporting. Real-time alerts were also generated for anomaly detection in the IoT dataset, using AWS SNS (Simple Notification Service) to notify users when anomalies were detected by the LSTM model.

4.2. Experimental Results

The experiments were evaluated based on several key performance metrics:
• Model Accuracy and Performance
For the credit card fraud detection dataset, the XGBoost model achieved an accuracy of 98.3%, with an AUC-ROC of 0.86. Training time was optimized through the use of SageMaker’s distributed infrastructure, with training completion in 15 minutes on an EC2 instance with 8 GPUs. The model's inference time on new data batches averaged 500 milliseconds, making it suitable for real-time detection as depicted in Figure 4.
Figure 4. Model accuracy and performance
AUC ROC Curve: ROC is a probability curve and AUC represents the degree or measure of separability. It tells how much the model is capable of distinguishing between classes. Higher the AUC, the better the model is at predicting 0 classes as 0 and 1 classes as 1.
By analogy, the higher the AUC, the better the model is at distinguishing between patients with the disease and no disease.
For the IoT sensor dataset, the LSTM model achieved 95% accuracy in anomaly detection. The use of Google Cloud TPU reduced the training time to 3 hours for the full dataset, compared to 12 hours on a CPU-based system. The model was able to detect anomalies with an average latency of 200 milliseconds when deployed in the cloud, allowing for near real-time analysis of sensor streams.
• Scalability
The framework demonstrated excellent scalability, particularly in the data ingestion and storage layers. When the IoT sensor dataset was expanded by 1 million additional readings, the AWS Kinesis stream continued to ingest data without any noticeable drop in performance. Similarly, Amazon S3 handled the expanded dataset size efficiently, with minimal delays in querying and loading data for model training.
We scaled the AWS EC2 instances from 4 GPUs to 8 GPUs during the XGBoost model training, reducing training time by 40%. The LSTM model on Google TPU showed linear scalability, maintaining training times as the dataset increased in size.
• Cost Efficiency
One of the critical aspects evaluated was cost efficiency. By using AWS Lambda for orchestration and serverless processing, we reduced unnecessary cloud costs. The use of spot instances on AWS EC2 for model training reduced costs by 30% compared to using on-demand instances. Similarly, AWS S3’s tiered storage model allowed us to optimize costs by moving less frequently accessed data to Infrequent Access storage tiers, resulting in a cost reduction of 20% for storage.
The total cloud cost for training and deploying the XGBoost model was $150, while the LSTM model on Google Cloud TPU incurred a cost of $250 for training and deployment over one week.
These costs are significantly lower than those incurred using traditional on-premises infrastructure, making the framework suitable for organizations with budget constraints.
• Real Time Processing
The framework demonstrated excellent performance in real-time processing, particularly with the IoT sensor dataset. The combination of AWS Lambda and Kinesis enabled near-instant ingestion and processing of data.
The LSTM model was able to detect anomalies within 200 milliseconds after data ingestion, enabling timely alerts and interventions in IoT-based monitoring systems.

4.3. Comparison with Existing Models

To evaluate the effectiveness of the proposed framework, we compared its performance against existing models in both traditional AI infrastructure (on-premises), standard cloud-based AI implementations and Google AutoML. We considered various metrics, including model accuracy, training time, cost-efficiency, scalability, and real-time processing capabilities.
We considered various metrics, including model accuracy, training time, cost-efficiency, scalability, and real-time processing capabilities.
The proposed framework outperforms both existing cloud and on-premises models across all metrics as shown in Table 1. The AI models in our framework achieved higher accuracy than those trained in traditional environments due to better resource allocation and optimized cloud infrastructure. The use of cloud-native AI services, like AWS SageMaker and Google Vertex AI, significantly reduced training time, making the framework more suitable for large datasets and time-sensitive applications. By utilizing serverless computing and spot instances, the proposed framework minimized cloud costs while maintaining high performance, showing a 30% cost reduction compared to existing models. The framework excels in real-time data analysis, offering lower latency than traditional cloud and on-premises solutions, a crucial advantage for industries requiring real-time insights, such as finance and IoT.

5. Conclusions

This research introduces an AI-driven data analysis framework that leverages cloud computing to improve scalability, cost-efficiency, and real-time processing. By integrating cloud-native services with advanced AI models, the framework addresses the challenges of large-scale data processing and model training. Experimental results show superior performance in key areas such as model accuracy, training time, cost savings (30% reduction), and real-time analytics.
The framework’s low-latency insights make it valuable for industries like IoT, finance, and healthcare. Future work will refine security, compliance, and scalability, positioning this framework as a foundation for future cloud-based AI innovation. This architecture can be used for the implementation of AI/ML solutions across the industries like Retail/Finance/Media/Telecom.

ACKNOWLEDGEMENTS

We would like to acknowledge Dollar General Corporation and University of North Carolina, Charlotte for providing guidance and help in this research work. We appreciate the continuous encouragement and provided lab work environment to complete this research.

References

[1]  N. Thakur, A. Singh, A.L. Sangal, Cloud services selection: A systematic review and future research directions, Computer Science Review 46 (2022) 100514. https://doi.org/10.1016/j.cosrev.2022.100514.
[2]  A. Belgacem, S. Mahmoudi, M. Kihl, Intelligent multi-agent reinforcement learning model for resources allocation in cloud computing, Journal of King Saud University - Computer and Information Sciences 34 (2022) 2391–2404. https://doi.org/10.1016/j.jksuci.2022.03.016.
[3]  Z. Zhou, L. Zhao, Cloud computing model for big data processing and performance optimization of multimedia communication, Computer Communications 160 (2020) 326–332. https://doi.org/10.1016/j.comcom.2020.06.015.
[4]  A. Fernandez, S. Garcia, F. Herrera, SMOTE for Learning from Imbalanced Data_ Progress and Challenges, Marking the 15-year Anniversary.pdf, Journal of Artificial Intelligence Research 61 (2018) 863–905.
[5]  J. Dean, G. Corrado, R. Monga, K. Chen, M. Devin, M. Mao, M. Ranzato, A. Senior, Large Scale Distributed Deep Networks, Advances in Neural Information Processing Systems 25 (2012) 1–9.
[6]  M. Bahrami, M. Singhal, The Role of Cloud Computing Architecture in Big Data, Studies in Big Data 8 (2015) 275–295. https://doi.org/10.1007/978-3-319-08254-7_13.
[7]  S.A. El-Seoud, H.F. El-Sofany, M. Abdelfattah, R. Mohamed, Big data and cloud computing: Trends and challenges, International Journal of Interactive Mobile Technologies 11 (2017) 34–52. https://doi.org/10.3991/ijim.v11i2.6561.
[8]  S.A. Bhat, N.F. Huang, Big Data and AI Revolution in Precision Agriculture: Survey and Challenges, IEEE Access 9 (2021) 110209–110222. https://doi.org/10.1109/ACCESS.2021.3102227.
[9]  H.K. Mistry, C. Mavani, A. Goswami, R. Patel, The Impact Of Cloud Computing And Ai On Industry Dynamics And Competition, Educational Administration: Theory and Practice 30 (2024).
[10]  C. Quinn, Future Trends and Emerging Technologies in AI-Driven Healthcare, Artificial Intelligence in Medicine (2024) 295–314. https://doi.org/10.1142/9789811284113_0018.
[11]  N. Ahmed, A. Abraham, Modeling Cloud Computing Risk Assessment Using Machine Learning, Advances in Intelligent Systems and Computing 334 (2015) 315–325. https://doi.org/10.1007/978-3-319-13572-4.
[12]  X. Wang, X. Xu, Q.Z. Sheng, Z. Wang, L. Yao, Novel artificial bee colony algorithms for QoS-aware service selection, IEEE Transactions on Services Computing 12 (2019) 247–261. https://doi.org/10.1109/TSC.2016.2612663.
[13]  I.A. Ansari, M. Pant, Quality assured and optimized image watermarking using artificial bee colony, International Journal of Systems Assurance Engineering and Management 9 (2018) 274–286. https://doi.org/10.1007/s13198-016-0568-2.
[14]  L.A. Tawalbeh, R. Mehmood, E. Benkhlifa, H. Song, Mobile Cloud Computing Model and Big Data Analysis for Healthcare Applications, IEEE Access 4 (2016) 6171–6180. https://doi.org/10.1109/ACCESS.2016.2613278.
[15]  N. Thakur, A.K. Sharma, Data Integrity Techniques in Cloud Computing: An Analysis, International Journal of Advanced Research in Computer Science and Software Engineering 7 (2017) 121. https://doi.org/10.23956/ijarcsse.v7i8.36.
[16]  N. Thakur, A. Singh, A.L. Sangal, Comparison of Multi-Criteria Decision-Making Techniques for Cloud Services Selection, Lecture Notes in Electrical Engineering 855 (2022) 669–682. https://doi.org/10.1007/978-981-16-8892-8_51.
[17]  N. Thakur, A.K. Sharma, DATA INTEGRITY CHECK IN CLOUD COMPUTING: A, International Journal of Computer Engineering and Applications XI (2017).
[18]  S. Kumar, W.M. Lim, U. Sivarajah, J. Kaur, Artificial Intelligence and Blockchain Integration in Business: Trends from a Bibliometric-Content Analysis, Information Systems Frontiers (2022). https://doi.org/10.1007/s10796-022-10279-0.
[19]  D.A. Johnson, M.M. Trivedi, Driving style recognition using a smartphone as a sensor platform, in: IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, 2011: pp. 1609–1615. https://doi.org/10.1109/ITSC.2011.6083078.
[20]  A. Muthanna, A.A. Ateya, A. Khakimov, I. Gudkova, A. Abuarqoub, K. Samouylov, A. Koucheryavy, Secure and reliable IoT networks using fog computing with software-defined networking and blockchain, Journal of Sensor and Actuator Networks 8 (2019). https://doi.org/10.3390/jsan8010015.
[21]  Y. Lecun, Y. Bengio, G. Hinton, Deep learning, Nature 521 (2015) 436–444. https://doi.org/10.1038/nature14539.
[22]  V.K. Damera, A. Nagesh, M. Nagaratna, Trust evaluation models for cloud computing, International Journal of Scientific and Technology Research 9 (2020) 1964–1971.
[23]  M. Adel Serhani, H.T. El-Kassabi, K. Shuaib, A.N. Navaz, B. Benatallah, A. Beheshti, Self-adapting cloud services orchestration for fulfilling intensive sensory data-driven IoT workflows, Future Generation Computer Systems 108 (2020) 583–597. https://doi.org/10.1016/j.future.2020.02.066.
[24]  Y. Chen, Y. Lu, L. Bulysheva, M.Y. Kataev, Applications of Blockchain in Industry 4.0: a Review, Information Systems Frontiers (2022). https://doi.org/10.1007/s10796-022-10248-7.
[25]  S. Ahmadi, Systematic Literature Review on Cloud Computing Security: Threats and Mitigation Strategies, Journal of Information Security 15 (2024) 148–167. https://doi.org/10.4236/jis.2024.152010.
[26]  S. Lins, K.D. Pandl, H. Teigeler, S. Thiebes, C. Bayer, A. Sunyaev, Artificial Intelligence as a Service, Business & Information Systems Engineering 63 (2021) 441–456. https://doi.org/10.1007/s12599-021-00708-w.
[27]  Maintained and managed Code reference GitHub link: https://github.com/RaviKcse08/datascience_projs/blob/main/CC_FraudDetection_dataSimulator.ipynb.