View: session overviewtalk overview
Registration will start on September, 16 at 8:00 at the desk nearby Room 048 (ground floor, on left from the main hall).
10:30 | Tracer: A Tool for Race Detection in Software Defined Network Models ABSTRACT. Software Defined Networking (SDN) has become a new paradigm in computer networking, introducing a decoupled architecture that separates the network into the data plane and the control plane. The control plane acts as the centralized brain, managing configuration updates and network management tasks, while the data plane handles traffic based on the configurations provided by the control plane. Given its asynchronous distributed nature, SDN can experience data races due to message passing between the control and data planes. This paper presents Tracer, a tool designed to automatically detect and explain the occurrence of data races in DyNetKAT SDN models. DyNetKAT is a formal framework for modeling and analyzing SDN behaviors, with robust operational semantics and a complete axiomatization implemented in Maude. Built on NetKAT, a language leveraging Kleene Algebra with Tests to express data plane forwarding behavior, DyNetKAT extends these capabilities by adding primitives for communication between the control and data planes. Tracer exploits the DyNetKAT axiomatization and enables race detection in SDNs based on Lamport vector clocks. Tracer is a publicly available tool. |
10:50 | PRESENTER: Georgios V. Pitsiladis ABSTRACT. The ecosystem of Privacy Calculus is a formal framework for privacy comprising (a) the Privacy Calculus, a Turing-complete language of message-exchanging processes based on the π-calculus, (b) a privacy policy language, and (c) a type checker that checks adherence of Privacy Calculus terms to privacy policies. BPMN is a standard for the graphical description of business processes which aims to be understandable by all business users, from those with no technical background to those implementing software. This paper presents how (a subset of) BPMN diagrams can be converted to Privacy Calculus terms, in the hope that it will serve as a small piece of larger workflows for building privacy-preserving software. The conversion is described mathematically in the paper, but has also been implemented as a software tool. |
11:40 | Complexity of Monomial Prediction in Cryptography and Machine Learning PRESENTER: Mahesh Rajasree ABSTRACT. In this paper, we focus on the monomial prediction problem in two settings: (1) Decide whether a particular monomial m is present in a composite function f:= f_r \circ f_{r-1} \circ \hdots f_0, where f_i are quadratic boolean functions, (2) Decide whether a particular monomial m is present in a composite function f:= f_r \circ f_{r-1} \circ \hdots f_0, where polynomials f_i are efficiently computable by Probabilistic Generating circuits over rationals. Probabilistic generating circuits (PGCs) are economical representations of multivariate probability generating polynomials (PGPs), which capture many tractable probabilistic models in machine learning. The first problem has a strong connection with the security of symmetric-key primitives. Dinur and Shamir proposed the cube attack for distinguishing a cryptographic primitive from a random function, which can be thought of as an efficient monomial prediction. In a general setting, over any large finite field or integers, monomial prediction is known to be NP-hard. Here, we show that in the quadratic setting, the problem is \oplus P-complete. \oplus P is an interesting complexity class that is not known to contain NP, however, it is believed to contain computationally hard problems. On the other hand, we also present several new zero-sum distinguishers for 5-round Ascon, which is one of the ten finalists for NIST light weight cryptography standardization competition. We show that the second problem is #P-complete. It is known that PGCs have efficient inference, i.e. given a monomial, one can efficiently output (which signifies the probability) its coefficient in the polynomial computed by the circuit. However, a composition of such functions makes the inference hard. Composition of probabilistic models and their efficient inference play a crucial role in the semantic contextualization and framework of uncertainty theories in graphical modelling. |
12:00 | On the Maximum Distance Sublattice Problem and Closest Vector Problem PRESENTER: Mahesh Rajasree ABSTRACT. In this paper, we introduced the Maximum Distance Sublattice Problem (MDSP). We observed that the problem of solving an instance of the Closest Vector Problem (CVP) in a lattice L is the same as solving an instance of MDSP in the dual lattice of L. We give an alternate reduction between the CVP and MDSP. This alternate reduction does not use the concept of dual lattice. |
12:20 | Abstract Continuation Semantics for a Biologically-Inspired Formalism ABSTRACT. We investigate the abstractness of a continuation semantics for a calculus inspired by DNA computing. This semantic investigation is given in the framework of complete metric spaces, and uses the weak abstractness criterion introduced by us in recent work. We prove that the denotational semantics designed with continuations is weakly abstract with respect to the operational semantics of our calculus which involves multiparty synchronization. We show that the expected concurrency laws are satisfied in this semantics for the calculus under investigation. |
12:40 | Optimizing Cold Start Performance in Serverless Computing Environments ABSTRACT. Serverless computing, also known as Function as a Service (FaaS), simplifies cloud application development by abstracting server management. However, cold start latency, which occurs when initializing the execution environment, poses significant challenges. This work proposes an Apache OpenWhisk-integrated dynamic caching and request routing system to mitigate cold start latency. By leveraging intelligent caching, the system aims to reduce response times for serverless functions. Comprehensive performance evaluations demonstrate that the proposed system reduced average response times by up to 7 times, while also ensuring efficient resource utilization. These findings provide practical insights and scalable solutions to enhance the efficiency and reliability of serverless architectures, particularly for latency-sensitive applications. |
14:00 | Novel Data Adaptation Techniques for Enhanced Lung Cancer Detection in CT Scans ABSTRACT. Lung cancer persists as a global leader in cancer-related deaths, highlighting the critical need for precise and efficient detection methods. This paper investigates the use of the Medical Segmentation Decathlon dataset to train neural networks for lung cancer segmentation in CT scans via semantic segmentation. We propose and evaluate four new data adaptation techniques specifically designed for this dataset, with each technique being assessed using U-Net-based architectures. Our approach incorporates a thorough exploratory data analysis to uncover the dataset's strengths and weaknesses, which in turn guided our data preprocessing and augmentation strategies. |
14:20 | Pre-Diagnosis of Autism in Children Using Distributed Learning PRESENTER: Alexandru Robert Vlasiu ABSTRACT. Autism Spectrum Disorder (ASD) is a multifaceted neurodevelopmental condition, exhibiting diverse symptoms related to social communication, behavior, and sensory process- ing. Diagnosing ASD in children necessitates a comprehensive assessment of their behavioral traits and the identification of crucial indicators, rendering the task intricate and nuanced. This study addresses the performance resulting from mixing machine learning algorithms like logistic regression, decision tree, Na ̈ıve Bayes, support vector machine, and random forests, to streamline the diagnostic process of ASD in pe- diatric populations, within the framework of a modified version of the Privacy-Preserving Ag- gregation of Teacher Ensembles (PATE) Frame- work. This approach facilitates the amalgama- tion of modest datasets from multiple practition- ers and an ensemble of independent distributed Machine Learning (ML) approaches while ensur- ing the preservation of data integrity through encryption mechanisms. Our findings highlight the efficacy of the PATE approach in automating the diagnosis in chil- dren, demonstrating enhanced sensitivity, speci- ficity, and accuracy while ensuring data privacy compared to individual techniques. This study contributes to the advancement of collabora- tive methods for predictive modeling in psychol- ogy, enabling practitioners with small datasets to combine their efforts effectively. |
14:40 | On applying GraphLime for trustworthy diabet prediction ABSTRACT. This paper aims to provide a comprehensive analysis of the benefits of employing GraphLIME (Local Interpretable Model Explanations for Graph Neural Networks) for reliable diabetes mellitus prediction. Our focus is on highlighting the advantages of integrating GraphLIME with a features attention-mechanism, compared to the standard pairing of deep learning neural networks with the original LIME explainability method. This system enabled us to develop an effective approach for identifying the most relevant features and applying the attention mechanism solely to those features. We conducted a detailed comparison of the performance metrics between the two approaches. By incorporating the attention mechanism, the model reached an accuracy of 92.6% in addressing the problem. The model’s performance is thoroughly illustrated, with results further assessed using the Receiver Operating Characteristic (ROC) curve. Applying this technique to a dataset of 768 patients with or without diabetes mellitus, we enhanced the model’s performance by over 18%. |
15:00 | How to Tackle Fake Medical Content with Clustering and Transformers PRESENTER: Radu Razvan Slavescu ABSTRACT. Spreading fake medical information / news online might threat the public health, e.g., by decreasing the vaccination rates or the level of trust in the healthcare process. The system presented in our paper aims to address this by determining in an automatic manner whether a text either contains or implies some claims known to be false. To this end, it relies on the Monant dataset, which comprises medical posts and claims whose veracity has been established by experts in the field. The dataset is used in two ways. First, the set of false claims aimed to be detected was obtained from it. These false claims are clustered, allowing the next steps to be performed against one cluster only, thus reducing the response time. Second, to detect a post's falsehood, this gets summarized using transformers, then the summary is compared similarity-wise with each element of the chosen cluster of false claims, using all-MiniLM-L6-v2 and a variation of BioBERT. A high similarity score is interpreted as the presence of the false claim in the summary. The results obtained from the experiments demonstrate, through manually checked examples, that the system manages to identify about 77% of the false claims within the text at hand (with an F1-value of 0.73). |
15:50 | On Projective Delineability PRESENTER: Lucas Michel ABSTRACT. We consider cylindrical algebraic decomposition (CAD) and the key concept of delineability which underpins CAD theory. We introduce the novel concept of projective delineability which is easier to guarantee computationally. We prove results about this which can allow reduced CAD computations. |
16:10 | Towards Verified Polynomial Factorisation ABSTRACT. Computer algebra systems are really good at factoring polynomials, i.e. writing f as a product of irreducible factors. It is relatively easy to verify that we have a factorisation, but verifying that these factors are irreducible is a much harder problem. This paper reports work-in-progress to do such verification in Lean. |
16:30 | First steps towards Computational Polynomials in Lean ABSTRACT. The proof assistant Lean has support for abstract polynomials, but this is not necessarily the same as support for computations with polynomials. Lean is also a functional programming language, so it should be possible to implement computational polynomials in Lean. It turns out not to be as easy as the naive author thought. |
16:50 | Newton's Polynomial Root-Finder Can Be Strengthened ABSTRACT. Our recent near-optimal polynomial root-finder at ACM-SIAM SODA 2024 enhances the power of the subdivision classical iterations, extends it to black box polynomials, given by an oracle (black box subroutine) for their evaluation rather than by coefficients, and promises to compete for user's choice. Restricted to a fixed Region of Interest on the complex domain the algorithm runs at the cost proportional to the number of roots in a small neighborhood of the Region. Polynomial root-finders based on functional iterations, in particular Newton's, have weaker formal support but compete empirically for approximation of all complex roots under proper initialization that involve polynomial coefficients. We accelerate Newton's and two other popular functional iterations by means of incorporation of the Fast Multipole method (FMM), devise fast black box algorithms for their initialization, and study combination of subdivision iterations with Newton's to enhancing their power for root-finding in a fixed complex region. Part of our study can be of independent technical interest for polynomial root-finding. |
15:50 | Frequency generation for a quantum current sensor in smart grids ABSTRACT. Integrating renewable energy sources, electric vehicles, and storage systems into power grids demands advanced control and monitoring systems. Precise current sensors are a critical component of these systems, essential for ensuring reliable and efficient electricity distribution. Quantum sensors, particularly those based on nitrogen vacancy (NV) centers in diamond, provide a promising solution for high-precision current measurement. This document introduces an innovative method for generating microwave signals using direct digital synthesis (DDS) and field-programmable gate arrays (FPGAs) to drive multiple signals. |
16:10 | PRESENTER: Peter Gloesekoetter ABSTRACT. We investigate the combined temperature and magnetic field dependent behavior of high NV-density microdiamonds in an all-optical frequency-domain based setup, utilizing the change in fluorescence lifetime. We use these frequency-domain data to train neural networks for the prediction of temperatures and magnetic fields. At zero magnetic field, we show the ability to predict temperatures in the range of 0°C to 100°C, with a standard deviation of 1.24°C. Furthermore, we show the ability to predict applied magnetic fields, while varying the ambient temperature, with a higher accuracy, compared to the approach of observing of the fluorescence intensity at a single excitation frequency. |
16:30 | Data-Driven Modeling of Photovoltaic Panels Using Neural Networks ABSTRACT. Photovoltaic (PV) systems are essential for the shift towards sustainable energy. Accurate performance modeling of these systems is vital for optimizing their efficiency and adapting to changing environmental conditions. Traditional analytical models often struggle with real-world complexities. This paper uses a neural networks (NNs) method to model PV system performance, offering superior accuracy and adaptability. By leveraging NNs, the model adopts a black-box methodology, bypassing the need for detailed knowledge of PV system complexities. The study utilizes historical performance data from two panels located in Sion, Switzerland, and Eugene, Oregon, in the United States. This approach captures the interdependencies between input parameters such as solar irradiance, temperature, and applied load and output variables such as voltage and current. Additionally, it accounts for key PV metrics including open circuit voltage, short circuit current, maximum power point current, and maximum power point voltage. The NN model is trained with a comprehensive dataset featuring high persistent excitation along with real measurements. Performance metrics such as root mean square error demonstrate that the data-driven NN model surpasses traditional methods, offering a more practical solution for PV system performance modeling. |
16:50 | Assessing Predictor Influence in LSTM Models for Enhanced Solar Energy Forecasting ABSTRACT. Accurate forecasting of solar energy production is highly important for an adequate integration of renewable energy into the power grid. This study explores the importance of various predictors for enhancing the accuracy of solar energy forecasting using two distinct Long Short-Term Memory models: one that relies solely on historical power production data and past exogenous weather data, and another that additionally incorporates future exogenous weather data. Various settings are tried to improve the results and relatively similar parameters are used by both, yielding more accurate results for the model that also uses information from the future. The results also demonstrate that predictor importance varies significantly between the two models. These findings suggest that different models benefit from distinct predictor combinations and incorporating exogenous data can significantly enhance forecasting performance. |
17:10 | Optimal decision tree design for near real-time control of battery energy storage systems ABSTRACT. The increasing numbers of distributed energy resources (DERs) incur new challenges for the energy supply due to the volatilities and uncertainties of renewable energies. To utilize the benefits of DERs, a combination with storage systems and intelligent controls are necessary. Commonly used control methods are based on Model Predictive Control (MPC) with forecast and optimization, which require computational capabilities at the DER on a level, on which the distribution system operator (DSO) has no access. This paper proposes a control algorithm based on data, which combines a model based and an artificial intelligence (AI) based approach to utilize benefits from both methods while compensating for their drawbacks. Mixed Integer Quadratic Programming (MIQP) is used to generate training and testing data for a decision tree (DT). The investigation of the optimal composition of the training data and the optimal architecture of the DT are the main focus of this paper. |
17:10 | Improving the Nested Markov Chain hyper-heuristic (NMHH) framework efficiency through sequential probability ratio testing ABSTRACT. This paper introduces sequential probability ratio testing to the hyper-heuristic design process. Hyper-heuristics often require many function evaluations to find an acceptable heuristic configuration for the given problem. The reduction of the number of function evaluations is proposed via sequential T-Testing. The new approach is compared to two state-of-the-art hyper-heuristics and six metaheuristics on twelve benchmark problems in the continuous domain. The results point to the effectiveness of the new approach. |
17:30 | An application of the Shepard operator in image reconstruction ABSTRACT. We propose an application of the Shepard operator combined with two radial basis functions in image processing. This method aims to reconstruct damaged black-and-white or color images, considering both a global and a local approach. In the construction of the Shepard operator, we use the inverse quadratic and the inverse multiquadric radial basis functions. |
17:50 | Generic high-speed design with low-area implementations of statistical operations based on an FPGA device PRESENTER: Mouhamad Chehaitly ABSTRACT. Artificial Intelligence has emerged as a transformative technology, revolutionizing numerous industries by enabling advanced automation, predictive analytics, and decisionmaking capabilities. For that Artificial Intelligence overruns many domains like telecommunication, smart manufacturing industry, autonomous machines, Automated Disease Diagnosis in Medical Imaging, defense, and others. On the other hand, the hardware implementation of Artificial Intelligence comes with certain challenges and constraints, especially in a critical area, which leverages machine learning algorithms and realtime data analysis to optimize production processes and improve overall efficiency. Statistical operations play a crucial role in various machine learning algorithms to understand, process data, or make predictions to optimize models. So, in this work, we developed a high-speed and low-area design and implemented statistical operations for image or signal processing using an FPGA Device. To enhance the performance, we develop different hardware architectures based on different levels of parallelism to process the statistical operations to compute the Mean, Variance, and RMS (Root Mean Square). These generic architectures work in parallel/pipeline architectures with and without memory. The proposed architectures implement an FPGA target (Intel/Altera Agilex 7: AGMH039R47A2E1V) using Altera Quartus prime pro edition version 23.4 and achieve an ultra-high throughput with low-area consumption compared to the state-of-art methods. For 480×640 image size, the mean calculation architecture involves 1498 logic registers, 1912 slice LUT, and just 29kbits memory and it operates at a maximum frequency of 406.5MHz. Additionally, for an 8×8 image size, we need 33 clock cycles to achieve the mean calculation and 33+1 clock cycles to complete the variance calculation, compared to other approaches that require more than 64 clock cycles. |