CLOUDCOM2024: 15TH IEEE INTENATIONAL CONFERENCE ON CLOUD COMPUTING TECHNOLOGY AND SCIENCE
PROGRAM FOR TUESDAY, DECEMBER 10TH
Days:
previous day
next day
all days

View: session overviewtalk overview

09:00-10:30 Session 7: Cloud Computing Applications
Chairs:
Zeyar Aung (Khalifa University of Science and Technology, Masdar Institute, UAE)
Lianying Zhao (Carleton University, Canada)
09:00
Huiyao Dong (ITMO University, Russia)
Igor Kotenko (St. Petersburg Federal Research Center of the Russian Academy of Sciences (SPC RAS), Russia)
Enhancing Security in EV Charging Systems: A Hybrid Detection and Mitigation Approach
PRESENTER: Huiyao Dong

ABSTRACT. TThe rapid proliferation of electric vehicles (EVs) necessitates advanced charging infrastructure, which is increasingly reliant on cloud-based technologies and the Internet of Things (IoT). However, these systems are vulnerable to cyberattacks that could have severe repercussions, including power grid failures. This paper addresses the security vulnerabilities inherent in the EV charging system. We propose a novel hybrid security solution tailored for cloud-based EV charging systems that integrates time-series-based regression models with classification algorithms to detect and mitigate both power consumption anomalies and network intrusions effectively. Our approach includes a comprehensive analysis to identify vulnerabilities and threats for cloud-based EV systems, a dual-model system for anomaly and intrusion detection, and a feedback-based threshold adjustment mechanism to assist overflow and anomaly identification. We provide a detailed analysis of data communication threats from a cloud perspective, design a robust security model, and evaluate various models to select the most effective ones for real-time security management. The finally decided models present excellent accuracy and practical value. Our findings contribute to enhancing the resilience of EV charging systems against cyber-physical threats, ensuring more reliable and secure operations.

09:30
Ayesha Ubaid (University of Technology Sydney, Australia)
Xiaojie Lin (University of Technology Sydney, Australia)
Farookh Khadeer Hussain (University of Technology Sydney, Australia)
SW Forecaster:An Intelligent Data-Driven Approach for Water Usage Demand Forecasting
PRESENTER: Ayesha Ubaid

ABSTRACT. Short-term water usage demand forecasting is crucial for improving water management and controlling water supply for residential and industrial purposes. Several studies have depicted the usefulness of water usage demand forecasting in making smart cities sustainable. However, the real-world translation of such forecasting systems still needs to be discussed. In this research, we have developed +10 days short-term demand forecasting framework that utilizes the existing state-of-the-art statistical and deep learning models. The designed framework has been mapped to the utility company’s legacy water usage demand forecasting process to promote digitization and sustainability. The experimental results demonstrate the successful integration of the data-driven system into the existing system with improved forecast accuracy.

10:00
Muhammad Sharshar (Khalifa University, UAE)
Kareem Elzeky (Mohamed bin Zayed University of Artificial Intelligence, UAE)
Menna Othman (Egypt Japan University of Science & Technology, Egypt)
Seifaldin Khaled (Egypt Japan University of Science & Technology, Egypt)
Hosam Moubarak (Egypt Japan University of Science & Technology, Egypt)
Walid Gomaa (Egypt Japan University of Science & Technology, Egypt)
Blockchain-Based Framework for Stock Market Operations: IPOs, Trading, and Dividend Distribution

ABSTRACT. This study examines blockchain technology's potential to enhance the stock market by offering a decentralized alternative to the traditional system. The study explores the challenges of the current traditional stock market, such as long settlement times, weak transparency, dependency on centralized clearing houses, and heightened fees. The proposed novel system leverages blockchain technology to create a fully decentralized stock market encompassing all fundamental operations from Initial Public Offerings (IPOs) to stock trading and extending to distributing profits or dividends. It connects all market participants through an Ethereum consortium blockchain fuelled by smart contracts. This approach replaces traditional intermediaries like brokers with automated regulations, ensuring more rapid transaction settlements, reduced operational errors and fees, and strengthened security measures. The proposed blockchain-based system, consisting of four smart contracts for IPO, stock exchange, dividend distribution, and participant management, demonstrates remarkable performance, showing promise in efficiently handling the current load of the largest stock markets like the London Stock Exchange (LSE). Through theoretical and practical testing, the system achieved a remarkable throughput of 120 transactions per second and latency or finality of just 4 seconds, underscoring its capability to significantly enhance the efficiency of stock market operations.

10:30-11:00Morning Coffee Break

Service Area of Lecture Hall 1

12:00-13:30Lunch Break

Service Area of Lecture Hall 1

13:30-15:00 Session 9: Microservices and Containerization
Chairs:
Andreas Henschel (Khalifa University, UAE)
Cheng-Hsin Hsu (National Tsing Hua University, China)
13:30
Carmine Colarusso (University of Sannio, Italy)
Ida Falco (University of Sannio, Italy)
Eugenio Zimeo (University of Sannio, Italy)
A greedy data-anchored placement of microservices in federated clouds

ABSTRACT. In a multiple Cloud environment, the placement of execution environments is crucial and may be subject to data constraints. Data could be anchored to some environments due to regulatory compliance, data sovereignty issues, or performance optimization. Consequently, applications and microservices must be designed to operate efficiently with these constraints. This enforces specific placement strategies to obtain good performances and scalability. In this paper, we propose a technique to enforce a constrained data-centric deployment placement in a federated multi-cloud environment. This algorithm analyzes a graph model of microservices interaction, considering communication with data storage and adopting ``anchors" for implementing a data-centric placement strategy. The results show how a better placement based on data position in multiple Clouds improves performance in terms of overall system response time. This also allows microservices to offload near data sources for multi-cloud environments, improving overall system performance without violating data movement constraints.

14:00
Dasith Edirisinghe (University of Moratuwa, Sri Lanka)
Kavinda Rajapakse (University of Moratuwa, Sri Lanka)
Pasindu Abeysinghe (University of Moratuwa, Sri Lanka)
Sunimal Rathnayake (University of Moratuwa, Sri Lanka)
SpotKube: Cost-Optimal Microservices Deployment with Cluster Autoscaling and Spot Pricing

ABSTRACT. Microservices architecture, known for its agility and efficiency, is ideal for cloud-based software development and deployment. Often paired with containers and orchestration systems, it benefits from tools that automate resource management. Cloud platforms like AWS offer transient pricing options, such as Spot Pricing, which can significantly reduce costs. However, the dynamic demand and abrupt termination of spot VMs pose challenges. Containerization and orchestration systems can optimize microservices deployment costs using spot pricing while meeting consumer goals. We introduce SpotKube, an open-source, Kubernetes-based solution that uses a genetic algorithm for cost optimization. SpotKube autoscales clusters for microservices applications on public clouds with spot pricing, analyzing application characteristics to recommend optimal resource allocation. It features an elastic cluster autoscaler that ensures cost- effective deployments, maintains performance requirements, and handles node terminations gracefully, minimizing system availability impact. Evaluations with real public cloud setups show SpotKube’s effectiveness compared to alternative optimization strategies.

14:30
Yuki Asakura (Kyushu Institute of Technology, Japan)
Kenichi Kourai (Kyushu Institute of Technology, Japan)
An Efficient State-Saving Mechanism for Out-of-band Container Migration
PRESENTER: Kenichi Kourai

ABSTRACT. Many clouds provide containers as lightweight virtualized environments inside virtual machines (VMs). Containers can be migrated between source and destination VMs for various reasons such as load balancing. However, the performance of container migration is largely degraded by the load and virtualization overhead of VMs because the migration mechanism runs inside VMs. Conversely, the performance of containers is largely affected by the load of container migration. This paper proposes OVmigrate to enable out-of-band container migration, which migrates a container running inside a VM from the outside of the VM. OVmigrate analyzes and obtains the states of a container in the memory of the source VM using a technique called VM introspection. It enables the migration mechanism outside a VM to independently save the states of a container running inside the VM. We have implemented OVmigrate for Linux and KVM and compared the performance of state saving with the existing tool called CRIU running inside the VM.

15:00-16:00Afternoon Coffee Break

Service Area of Lecture Hall 1

15:00-16:00 Session 10: Poster Session
Chair:
Kamal Taha (Khalifa University, UAE)
Location: The Spine
Truong Thanh Le (Department of Informatics, University of Oslo, Norway)
Amir Taherkordi (Department of Informatics, University of Oslo, Norway)
Frank Eliassen (Department of Informatics, University of Oslo, Norway)
Peiyuan Guan (Department of Informatics, University of Oslo, Norway)
Optimal Distribution of ML Models over Edge for Applications with High Input Frequency
PRESENTER: Truong Thanh Le

ABSTRACT. The rise of more complex and sizeable Machine Learning (ML) models challenges traditional cloud computing models with respect to the high volume of incoming data which results in increased bandwidth usage and network congestion, as well as delays in inference. Such ML models are being rapidly developed thanks to advances in computing platforms and the real-time computing demands of ML-driven applications such as autonomous vehicles and video processing. To mitigate these challenges, ML model distribution and inference offloading to computing devices close to data sources have been explored, especially through \textit{partitioning} the models across the IoT-Edge-Cloud continuum. Existing efforts in this area have not successfully mastered the fully automatic and efficient determination of optimal partition points, along with the incorporation of Early-Exit layers which enable early termination of model inference at earlier stages when possible. In this paper, we introduce a novel partitioning algorithm designed to distribute ML models across edge devices with the goal of \textit{reducing response time when facing high-rate input data streams}. Our proposed approach leverages the principles of Dynamic Programming to determine optimal partition points and establish appropriate exit thresholds for Early-Exit layers. Our evaluation results reveal that, in the context of continuous, high-rate input data, our method consistently lowers the maximum round-trip time for processing inference requests compared to state-of-the-art methods such as NeuroSurgeon and Genetic.

Antonio Maciá-Lillo (Department of Computer Science and Technology, University of Alicante, Spain)
Higinio Mora (Department of Computer Science and Technology, University of Alicante, Spain)
Antonio Jimeno-Morenilla (Department of Computer Science and Technology, University of Alicante, Spain)
Tamai Ramírez (Department of Computer Science and Technology, University of Alicante, Spain)
Towards Abstraction of Heterogeneous Accelerators for HPC/AI Tasks in the Cloud

ABSTRACT. In the context of modern cloud computing, the integration of specialized hardware accelerators, including Graphics Processing Units (GPUs), Tensor Processing Units (TPUs), and Field-Programmable Gate Arrays (FPGAs), is pivotal for achieving high-performance computing (HPC) and supporting the intensive computational demands of artificial intelligence (AI) and machine learning (ML) workloads. Extensive work has been done in virtualization technology to support the use of these devices in cloud architectures. However, current research is focused on concrete technologies and devices. HPC and AI work benefit from the use of different accelerator devices. There is the need to generalize and account for the heterogeneity of devices and vendors. This paper proposes a conceptual model for a heterogeneous cloud architecture that schedules HPC and AI tasks across various accelerator devices. We propose the ``running rounds'' (RR) indicator, an analytical measure used in previous work, generalized for multiple GPU programming technologies, adapted for Intel GPUs, to predict the performance of jobs on different sets of resources. The efficacy of the model is demonstrated through experiments that validate the capability of the RR indicator to approximate performance, although with noted discrepancies that highlight the need for further refinement. The study underscores the potential and challenges of heterogeneous cloud environments.

Saman Akbari (Technische Universität Berlin, Germany)
Manfred Hauswirth (Technische Universität Berlin & Fraunhofer Institute for Open Communication Systems, Germany)
Sampling in Cloud Benchmarking: A Critical Review and Methodological Guidelines
PRESENTER: Saman Akbari

ABSTRACT. Cloud benchmarks suffer from performance fluctuations caused by resource contention, network latency, hardware heterogeneity, and other factors along with decisions taken in the benchmark design. Particularly, the sampling strategy that benchmark designers employ may significantly influence benchmark results. Despite this well-known fact, no systematic approach has been devised so far to make sampling results comparable and guide benchmark designers in choosing their sampling strategy for use within benchmarks. To identify systematic problems, we critically review sampling in recent cloud computing research. Our analysis identifies concerning trends: (i) a high prevalence of non-probability sampling, (ii) over-reliance on a single benchmark, and (iii) restricted access to samples. To address these issues, we propose methodological guidelines for sampling in cloud benchmarking. We hope that our work contributes to improving the generalizability, reproducibility, and reliability of cloud computing research.

Alexis Brodeur (École de Technologie Supérieure, Canada)
Guillaume Tassotti (École de Technologie Supérieure, Canada)
Amine Boukthouta (Ericsson, Canada)
Abdeljouad Necir Medakene (École de Technologie Supérieure, Canada)
Abdelouahed Gherbi (École de Technologie Supérieure, Canada)
tBPF: Testing Berkeley Packet Filter Programs Using User Mode Linux
PRESENTER: Alexis Brodeur

ABSTRACT. The Berkeley Packet Filter (BPF) is increasingly used for a variety of use cases including auditing, security, monitoring, networking, etc. However, BPF lacks improved and effective tools, which makes the integration of programs written for BPF quite hard in current automated testing solutions and continue integration pipelines. We present tBPF, a library for integration testing of BPF programs that allows automated testing of arbitrary programs in a kernel agnostic manner without superuser privileges. We show that our solution can be integrated in existing development workflows and pipelines to enable reproducible testing and auditing of BPF programs. In this article, we describe our approach and compare it against other approaches in relevant work and literature.

Saira Musa (McMaster University, Canada)
Douglas G. Down (McMaster University, Canada)
Cost-Effective Cloud Resource Provisioning Using Linear Regression
PRESENTER: Saira Musa

ABSTRACT. In the era of cloud computing, accessing virtual computing resources has become increasingly convenient for users to meet their demands. Cloud providers offer two primary payment plans for virtual resource provisioning: reservation and on-demand. The reservation plan requires users to reserve resources and pay upfront, making it more cost-effective for long-term requirements despite demand uncertainty. Conversely, the on-demand plan charges based on actual resource usage, making it more expensive but suitable for short-term needs. Efficient resource provisioning is crucial to balance user demands and costs, as inefficient provisioning can lead to high costs. A key challenge is determining the optimal number of resources to reserve to accommodate unpredictable demands while minimizing costs. This paper addresses the resource reservation problem in cloud environments by focusing on the optimal reservation of virtual machines (VMs). We propose a linear regression approach that fits a linear function to features such as past demands and previously reserved instances that are still available to determine the quantity of VMs to reserve. Our model assigns specific weights to these features to predict reserved instances, minimizing the overall cost, including both the expense of reserving resources and renting additional on-demand resources as needed. Our evaluation, based on real standard workload traces, demonstrates the effectiveness of our approach in achieving cost-efficient resource provisioning, reducing the total cost by efficiently balancing reserved and on-demand resources.

Chanh Nguyen (Umeå University, Sweden)
Monowar Bhuyan (Umeå University, Sweden)
Erik Elmroth (Umeå University and Elastisys, Sweden)
Enhancing Machine Learning Performance in Dynamic Cloud Environments with Auto-Adaptive Models
PRESENTER: Chanh Nguyen

ABSTRACT. In this paper, we propose an auto-adaptive ML model approach to mitigate the negative impact of data drift on model performance in cloud systems, leveraging the principles of generalization bounds and domain adaptation. Our method builds a knowledge base of time series data batches from historical cloud operational data, each linked to a trained ML model and clustered using the HDBSCAN algorithm. When performance degradation is detected, the system matches the latest inference data to a similar batch using Dynamic Time Warping (DTW) and applies the matched model's hyperparameters to the new data stream.

Extensive experiments on two real-world cloud system traces show that the proposed approach maintains high model accuracy while minimizing retraining overhead. In scenarios with frequent data drift, such as the Wikipedia trace, our framework reduces retraining overhead by 22.9% compared to drift detection-based retraining and by 97% compared to incremental retraining. In more stable environments, like the Google cluster trace, retraining costs are reduced by 96.3% and 88.9%, respectively. Scalability tests with large data inputs further show that the proposed approach efficiently handles large-scale data within a limited timeframe, making it well-suited for dynamic cloud environments.

16:00-18:00 Session 11: Special Session: Digital Health: What Do the Doctors Say?
Chair:
Leila Ismail (United Arab Emirates University, UAE)
16:00
Leila Ismail (United Arab Emirates University, UAE)
Opening Remarks: Reimagining Healthcare with Unprecedented Technological Advancements
16:10
Shahrukh Hashmi (Medical Research & Development Division, Department of Health, Abu Dhabi, UAE)
Regulatory Approvals for Software as a Medical Device
16:30
Moein A B Khan (College of Medicine and Health Sciences, Department of Family Medicine, UAE University, UAE)
Critical Health Care Challenges in Primary Health Care
16:50
Andreas Henschel (Khalifa University, UAE)
Feature Engineering Meets Immunology: Harnessing Clinical and Genetic Data with Machine Learning to Unravel Disease Mechanisms
17:10
Sreejith Balasubramanian (Middlesex University, UAE)
Assessing Readiness for Blockchain Adoption: Insights from a Healthcare Case Study
17:30
Leila Ismail (United Arab Emirates University, UAE)
Discussion and Closing Remarks