COHORT.4

2022 – 2026

Modal Title

Aaron Dees

Student

Project Title

A machine learning approach to women+ centred health across the lifecycle

Project Description

Historically women+ have been significantly underrepresented with regards to medical research (Merone, Tsey, Darren, Nagle, 2022). This is seen in both policies relating to women+ health, research in clinical trials but also the presentation and management of conditions in clinical settings (Merone, Tsey, Darren, Nagle, 2022)( Maucais-Jarvis, Merz et al., 2020). For example in 1977 the Food and Drug Administration in the US released policy guidlines for clinical trials in General Considerations for the clinical Evaluation of Drugs which was based on data that excluded all ‘’Women of Childbearing Potential’’ from phase 1 clinical trials, regardless if they were on birth control, single or if their husband had a vasectomy (U.S. FDA, 1977). It wasn’t until 1993 that the FDA revised its publication from 1977 allowing women+ to be included in all stages of clinical trials if it met certain criteria (U.S. FDA, 1993). This was done as there was a “growing concern that the drug development process does not produce adequate information about the effects of drugs on women’’(U.S. FDA, 1993).

In addition to the exclusion of women+ from clinical trials and health data collection processes, the presentation and prevalence of health conditions can vary according to gender ( Maucais-Jarvis, Merz et al., 2020) but this is not always considered in clinical management of female patients. For example, clinical evidence shows that women+ are half as likely to receive interventional medicine for coronary artery disease when compared to their male counterparts (Weisz, Gusmano, Rodwin, 2004).

Although there is clear evidence of exclusion and bias in women+ healthcare, one must first know where the deficiencies and bias lie within particular conditions to be able to appropriately address them. Machine learning has the potential to make a vast impact on women+ centered health by analysing health related data of various conditions, identifing these key defiencies and biases and addressing these key deficiencies to develop a more appropriate approach to the management of these conditions.

Research Aim: The aim of this research is to explore the deficiencies in the management of women+’s health conditions from presentation to diagnosis and treatment across the lifecycle and investigate issues of bias and exclusion.

Objectives: The first steps in this research will involve a qualitative study to explore the key deficiencies in the management of women+’s health with key experts in the areas of health (GP’s and Pharmacists) and health related policy makers. The data collected will be analysed using thematic analysis (Braun and Clarke, 2006) and the results will then directly inform subsequent data driven explorations in particular health conditions where deficiencies and biases have been identified.

Data Exploration – Some of the main sources of data initially identified include:

  • The Autoimmune Association {Charity/Research organisation}
  • DAISy PCOS {Research organisation}
  • SWAN datasets {Open Source Data }

Statistical Analysis of data from stage 2 to determine critical factors that have the largest impact for health outcomes for women+

Modal Title

Ajay Kumar M

Student

Project Title

Low power DNN Processor for Embedded AI in IoT sensors

Project Description

Deep neural networks (DNNs), which are loosely modelled after the human brain, have been shown to be
remarkably successful in recognizing and interpreting patterns in many applications, such as image and video
processing. However, due to high computational complexity, power, and resource requirements, DNNs are not
well explored for low power applications such as Internet of Things (IoT) sensors, wearable healthcare, etc.
IoT devices have stringent energy and resource constraints and, in many cases, deal with one-dimensional
time series data.

This project will investigate DNN hardware accelerator architectures for low-power edge implementation. The
project will address the challenges of implementing sparse DNN models in energy-efficient IoT edge sensors.
The DNN accelerator hardware developed will be based on a RISC-V CPU and will be scalable,
programmable, and easily adaptable to different DNN model types. The accelerator developed will be
demonstrated in an FPGA (or ASIC) for a wearable biomedical application such as heartbeat classification

Modal Title

Alexander Victor Okhuese

Student

Project Title

Developing and enabling edge analytics for Internet of Medical Things

Project Description

This project seeks to investigate and develop robust and resource-based edged algorithms for applications in the Internet of Things and the development of intelligent applications of real-time data-enhanced solutions in the e-Health domains.

Machine learning (ML) and deep learning (DL) are promising tools for developing intelligent applications for various domains, especially in the e-health domain. Unfortunately, these algorithms are typically very resource-intensive and require a lot of computational resources to deploy and test. The Internet of Medical Things (IoMT) is also a resource-constrained ecosystem, despite its evolution and applications in  recent times. Meanwhile, due to the ground-breaking deployment of 5G technology, the challenge of resource constraint seems to be solvable as new improved algorithms seek to improve connectivity, cloud-based storage, and extend the computability of machines while expanding the applicability of mobile IoMT (West, 2016; Patel et al., 2017). Evidently, there are emerging applications of edge IoMT devices in care delivery, as connected medicine seems to provide advanced dynamics for obtaining quality care through imaging improvement, diagnostics, and treatment of all types of complex health challenges.

Therefore, this project will be focused on building innovative solutions to design compact and lightweight machine learning and deep learning algorithms to fit, train, and deploy these algorithms on resource-constrained edge devices for tracking, early detection, and diagnosis of diseases without compromising their performance and accuracy. Considering the nature of tracking that is required for the device to perform and the eventual detection or prediction of the possible presence of diseases in the body, distributed and federated learning techniques will be used over multi-modal data sets (e.g. sensor data streams, historical datasets) to provide robust and accurate Artificial Intelligence (AI) models for application in tracking, detecting, and possible enhancement of the different chemotherapy stages for affected patients.

To deploy edged IoMT devices, such as clinical wearables and remote sensors, as well as many other devices that monitor and electronically transmit medical data, such as vital signs, physical activity, personal safety, and medication adherence, the project specifically aims to develop robust algorithms based on big data analytics, learning, and intelligence. Smart objects are now the ideal building blocks for the creation of cyber-physical smart ubiquitous frameworks thanks to the IoMT. With bright technological, economic, and societal prospects, the IoMT revolution is redefining

contemporary healthcare (Islam et al., 2015). These tools will offer diagnosis and treatment options in telemedicine that have never been seen before, all while providing high-quality care at a reasonable cost for patients.

In summary, this project seeks to explore the use of knowledge graph generation and explainable AI techniques over IoMT data for the tracking, detection, and diagnosis of diseases in the human body.

Modal Title

Alexandria Mulligan

Student

Project Title

ROP-StuDIO

Project Description

ROP-StuDIO is a research project centered around medical applications of machine learning regarding retinopathy of prematurity (ROP). At present, ROP is the leading cause of preventable childhood blindness and impacts premature newborns. The more premature a birth, the greater the risk of ROP and the increased risk medical scanning induces on the health of the child. While sometimes ROP is corrected naturally other times it is severe and treatment is required. Current ROP treatment is able to save a part vision in a child that would otherwise go blind. However, it is critical for ROP screening methods to predict the risk level of each newborn to reduce limiting sight on a child that may fully recover or missing treating another child who could have had some of their vision saved end up with no vision. In addition ROP screening methods require a specialised medical practitioner who is not always available in remote locations and the methods themselves are invasive for critical state premature births.

Research Aim:

This project aims to explore and address two research questions. The first, “can machine learning applications use clinical data of premature babies and their mothers to map risk assessment for infants developing ROP?”

Furthermore, the second question for exploration is, “can deep learning approaches be used to track ROP development using retinal fundus images?” The exploration of this question includes developing a set of guidelines around how to acquire optimal retinal scans for analysis. This secondary outcome intends to decrease the number of images and scans necessary for each baby.

Research Scope and Objectives:

The scope of this project is agile with seven main sprints or deliverable packages. The first is achieving domain understanding using a literature review and interacting with domain experts. One of these experts is the internal advisory supervisor for this project who has access to a substantial image dataset at Cork University Hospital. The second sprint focuses on this dataset and requires more domain training to understand the rental scans in the context of ROP. The third focus of the research project is to clean the data and perform necessary image pre-processing for the fourth sprint which seeks to develop a deep learning model for image evaluation. This fourth work package’s discoveries will be prepared for publication.

The fifth sprint is model development and testing for clinical data of the baby and mothers. This data provides a possible challenge in acquiring and will need to be applied for. The outcome of this model should identify premature babies as high risk or low risk for ROP due to the clinical data building off the current practice of using gestational age and weight as risk identifiers. The feature behaviours, and model developed in this stage will also be prepared for publication.

Work package six for this project evaluates the models from package four and five alongside clinical experts. The outcome of this is to determine if the models hold results that are acceptable in clinical practice. The final package of this project is thesis write up and submission combining the results and discoveries from all prior work packages.

Modal Title

Amina Said

Student

Project Title

Validation and Quantification of optical properties of water using machine learning

Project Description

Optical properties of water provide information about the condition of water bodies which can also give information on water pollution. Various water quality parameters such as the total suspended matter (TSM), chlorophyll and colour dissolved organic matter (CDOM) concentrations were traditionally based on estimations from in situ data. This method is expensive, time-consuming and provided local coverage. However, technology has been progressively improving and advancing. More remote sensing data and platforms are now freely and easily accessible. Water quality can now be extracted from the satellite imageries with relatively good spatial resolution, data is more frequently collected at a global coverage and at a good accuracy. A strong positive correlation coexists between the in situ data and satellite imageries on water properties. Though, machine learning can assist in further improving the accuracy and automate the process by using trained models. The size of dataset tends to influence the accuracy of remote sensing results. Training a model in machine learning requires a large dataset which eventually improves the accuracy of remote sensing. The process mainly involves preparing the data for training in a model. A suitable machine learning model is selected and model is fit into the data. It involves learning the pattern of the data and using the pattern to make predictions. Next, the model is evaluated to assess the accuracy and precision of the model. If the accuracy of the model is low, tuning and experimentation is carried out. Which improves the accuracy of the model by manipulating the hyperparameters and comparing accuracy of the model from the various hyperparameters used. The model is saved once the required accuracy is acquired. This can then be reloaded and reused for making predictions and classification. Which can assist in making prediction on water quality parameters, classification of satellite imageries into respective feature classes or detect specific features in an imagery. Thus, machine learning assists in predictions of both in situ and satellite data, classification and feature detection which ultimately saves time, money and improves the accuracy. However, various challenges are expected in this research project. Such as finding satellite imageries with good spatial resolution and with less than 10 percent of cloud cover. Moreover, training the model and improving its accuracy can be time-consuming. Despite the challenges, the research project will provide an insight into the potential of remote sensing and machine learning in extracting the water optical properties

Modal Title

Anam Hashmi

Student

Project Title

Interpretability and Visualization of Deep Learning Models in Magnetic Resonance Imaging and Radiology

Project Description

Deep learning approaches are achieving state-of-the-art results in Magnetic Resonance Imaging (MRI) and Radiology (Computed Tomography, X-ray) on such downstream tasks as image classification and image segmentation. However, they still suffer from a lack of human interpretability critical for increasing understanding of the methods’ operation, enabling clinical translation and gaining clinician trust. Therefore, it is crucial to provide clear and interpretable rationale for model decisions.

In order to understand model predictions, there are algorithms that can be applied to a fully trained deep learning model and produce heatmaps highlighting portions of the image that contributed to the model predictions. These algorithms are (but are not limited to) Guided-Backpropagation, Class Activation Mapping (CAM), Grad-CAM, Grad-CAM++ and Score-CAM.

The recent COVID‐19 pandemic made clear that rapid clinical translation of Artificial Intelligence (AI) systems may be critically important for assisting treatment‐related decisions and improving patient outcomes. Recently, there has been a particular interest in applying the above algorithms to interpret COVID-19 Chest X-rays and assist clinicians in making the correct diagnosis.

We will be working with the open source MRI, CT and X-ray public datasets and will initially investigate:

  1. How different Convolutional Neural Networks (CNNs) architectures contribute to the generation of reliable heatmaps. Here, we will focus on both custom CNNs and established pre-trained transfer learning based architectures such as VGG, ResNet, Inception, etc.
  2. Weight initialization in CNNs has been also shown (and is in line with our experience) to be important to the performance of the interpretability algorithms and we will investigate different weight initialization protocols. We will attempt to answer the research question why the initialization with ImageNet weights that are completely unrelated to the medical imaging domain leads to the astonishing performance of some interpretability algorithms on medical imaging datasets.
  3. Most works in the literature are focused on the interpretability methods for the classification models but there are some limited works that also examine the interpretability of segmentation networks. We will examine both downstream tasks and for this purpose we will be working with the 2D T1-weighted CE-MRI dataset compiled by Cheng et al. as this dataset contains both classification and segmentation labels.
  4. Therefore, this study will also consider weakly supervised image segmentation of the Cheng et al. dataset using class-specific heatmaps, possibly Conditional Random Fields (CRF), and thresholding techniques.
  5. Many interpretability algorithms are applied to the last convolutional layer of CNNs and therefore the resulting heatmaps are coarse. We will investigate how to leverage the intermediate layers for the generation of finer heatmaps.

To summarize, this project is expected to identify best performing approaches for providing explainable and interpretable AI output and discuss their advantages and disadvantages for the MRI and Radiology datasets considered in this study. Further, we expect to provide recommendations for appropriate incorporation of these techniques to improve deep learning models and evaluate their performance.

Modal Title

Andrea Heaney

Student

Project Title

A machine learning approach to women+ centred health across the lifecycle

Project Description

Historically women+ have been significantly underrepresented with regards to medical research (Merone, Tsey, Darren, Nagle, 2022). This is seen in both policies relating to women+ health, research in clinical trials but also the presentation and management of conditions in clinical settings (Merone, Tsey, Darren, Nagle, 2022)( Maucais-Jarvis, Merz et al., 2020). For example in 1977 the Food and Drug Administration in the US released policy guidlines for clinical trials in General Considerations for the clinical Evaluation of Drugs which was based on data that excluded all ‘’Women of Childbearing Potential’’ from phase 1 clinical trials, regardless if they were on birth control, single or if their husband had a vasectomy (U.S. FDA, 1977). It wasn’t until 1993 that the FDA revised its publication from 1977 allowing women+ to be included in all stages of clinical trials if it met certain criteria (U.S. FDA, 1993). This was done as there was a “growing concern that the drug development process does not produce adequate information about the effects of drugs on women’’(U.S. FDA, 1993).

In addition to the exclusion of women+ from clinical trials and health data collection processes, the presentation and prevalence of health conditions can vary according to gender ( Maucais-Jarvis, Merz et al., 2020) but this is not always considered in clinical management of female patients. For example, clinical evidence shows that women+ are half as likely to receive interventional medicine for coronary artery disease when compared to their male counterparts (Weisz, Gusmano, Rodwin, 2004).

Although there is clear evidence of exclusion and bias in women+ healthcare, one must first know where the deficiencies and bias lie within particular conditions to be able to appropriately address them. Machine learning has the potential to make a vast impact on women+ centered health by analysing health related data of various conditions, identifing these key defiencies and biases and addressing these key deficiencies to develop a more appropriate approach to the management of these conditions.

Research Aim: The aim of this research is to explore the deficiencies in the management of women+’s health conditions from presentation to diagnosis and treatment across the lifecycle and investigate issues of bias and exclusion.

Objectives: The first steps in this research will involve a qualitative study to explore the key deficiencies in the management of women+’s health with key experts in the areas of health (GP’s and Pharmacists) and health related policy makers. The data collected will be analysed using thematic analysis (Braun and Clarke, 2006) and the results will then directly inform subsequent data driven explorations in particular health conditions where deficiencies and biases have been identified.

Data Exploration – Some of the main sources of data initially identified include:

  • The Autoimmune Association {Charity/Research organisation}
  • DAISy PCOS {Research organisation}
  • SWAN datasets {Open Source Data }

Statistical Analysis of data from stage 2 to determine critical factors that have the largest impact for health outcomes for women+

Modal Title

Asad Ullah

Student

Project Title

Intermediate Speech Representations for Low Resource Speech Models

Project Description

Feature extraction and intermediate speech representation are important components for speech processing tasks. Many different approaches exist, e.g. methods such as wav2vec2, Mockingjay, TERA, autoregressive methods. Beyond taking the raw input wave, basic feature extractions (e.g. MFCC, log mel spectrogram etc.) are widely used in models. Some also create intermediate representations that are useful for self-supervised learning and to allow base models trained without labels to be fine-tuned and applied to a variety of different prediction tasks and target outputs. Understanding and explaining why these methods work well on some downstream tasks and but not others has not been well studied for different speech objectives such as phoneme recognition, speaker identification, speech recognition, language identification, spoken language understanding, speech translation, emotion recognition, voice conversion, speech synthesis etc.

This project will adapt state-of-the-art deep learning architecture to improve the existing speech representation methods. New methods emerging from the fields of computer vision (CV) and natural language processing (NLP) will be reviewed for cross-domain inspiration. Datasets will be sourced to fine-tune models with varying amounts of labelled data, This will inform the relationship between fine-tuning dataset size and the chosen representation, highlighting the potential for application to low resourced speech tasks. Better understanding of the relationship between the representations and the training data for the initial frozen model and fine-tuned models will help inform model and data choices across different classes of speech models.

Modal Title

Colm O'Donaghue

Student

Project Title

Investigating The Use of Machine Learning in Health Science

Project Description

The benefits in regular physical activity (PA) participation are well documented with irrefutable evidence of the effectiveness of regular PA in the prevention of several chronic diseases (e.g., cardiovascular disease, diabetes, obesity, depression, etc.) and premature death (Warburton et al., 2006). In Ireland, the Children’s Sport Participation and Physical Activity (CSPPA) study found just 17% of Irish children engaged in the recommended one hour per day of moderate to vigorous physical activity, which is a drop from the 19% recorded in 2010. (Woods et al., 2018). Therefore, it seems important to explore factors that influence PA participation and a means to measure impact.

Machine learning (ML) has become a common way to measure physical activity (Narayanan et al., 2020). Accelerometer data processing techniques based on pattern recognition have been shown to provide accurate predictions of physical activity type and more accurate assessments of physical activity intensity (Trost et al., 2012, 2018; Ellis et al., 2016). Nonetheless, the uptake of machine learning methods by physical activity researchers has been slow, in part due to the difficulties of implementation, and the consistent finding that models trained on accelerometer data from laboratory-based activity trials do not generalize well to free-living environments (Sasaki et al., 2016; Lyden et al., 2014; Bastian et al., 2015). Recently, it has been argued that ML has failed physical activity research in four important ways: a lack of benchmark data, priority in methods development, limited software integration and absence of training (Fuller, Ferber and Stanley, 2022). To improve the use of ML methods in physical activity research, it has been proposed that as a discipline, practitioners must use and publish benchmark datasets to allow for increased opensource methods development. Many datasets currently exist on health-related fitness, wellbeing, confidence and motivation towards PA amongst the Irish youth demographic, a recent example being the Moving Well-Being Well project, a national study which assessed a range of variables linked to PA participation in over 2,000 Irish children. One finding from the study highlighted the low levels of fundamental movement skills (FMS) mastery in Irish primary school children (Behan et al., 2019). FMS proficiency has shown positive associations with increased physical activity in both children and adolescents (Barnett et al., 2009; Lubans et al., 2010).

Current PA-related datasets are limited in terms of the analyses undertaken, due to a multitude of factors, a lack of training being one of them (Fuller, Ferber and Stanley, 2022). Given the size and complexities of these datasets, as well as the societal need to create a healthier Ireland, it would appear necessary that relevant ML techniques are explored in an effort to uncover further unknown patterns and correlations/disparities that exist between specific age groups and genders. This research aims to explore the use of ML in health science, specifically focusing upon children and youth with a further aim of utilizing ML in an appropriate manner towards PA to have a societal impact on Irish health.

Modal Title

Conor O'Sullivan

Student

Project Title

Monitoring Ireland’s coastal areas from satellite imagery using deep neural networks

Project Description

Monitoring coastal evolution linked to climate change is a difficult task. Coastlines and water bodies are impacted by climate changes and local weather conditions. It is essential to be able to monitor changes occurring along coastlines, because these changes impact populations living in coastal areas. In this PhD project, we intend to explore deep neural networks on open-access Copernicus Sentinel-2 satellite image data to develop machine learning methodologies to monitor climate change-induced coastal evolution along the Irish coast and other inland regions. This thesis involves forecasting Irish sea levels using Long short-term memory (LSTM)-based neural networks and classifying wetlands and other inland objects from satellite imagery.

Modal Title

Davide Serramazza

Student

Project Title

Explanation Methods for Multivariate Time Series Classification

Project Description

Multivariate time series classification is an important computational task arising in applications where data is recorded over time and over multiple channels. For example a mobile phone or a smartwatch can record the acceleration and orientation of a person’s motion, and these signals are recorded as multivariate time series. One can classify this data to understand and predict human movement and various conditions such as fitness levels. Classification alone is not enough in many applications. In most applications we need to classify but also to understand what the model learns (e.g. why a prediction was given, based on what information in the data?).

The main focus of this project is on understanding, developing and evaluating explanation methods tailored to Multivariate Time Series Classification (MTSC).

 

Some of the main research questions are:

  1. What are the current explanation methods for MTSC, and what are their strengths and limitations?
  2. How do we objectively evaluate and compare multiple explanation methods for MTSC, especially for scenarios where their outputs disagree (e.g. for saliency-based methods, different explanations provide very different saliency maps)?
  3. How can we integrate the insights gained from the explanation methods to improve the accuracy of the classification algorithms (eg an iterative optimisation approach)?
  4. How can we integrate the insights gained from the explanation methods to improve or reduce the input data (eg, by removing noisy data, selecting good features)?
Modal Title

Eanna Curran

Student

Project Title

Neural Network Architecture for Solving Geometric Optimization Problems

Project Description

Combinatorial optimization problems arise naturally in many areas of computer science and other disciplines, such as business analytics, operations research, bioinformatics and electronic commerce. Since many of these optimization problems are NP-hard, applications typically rely on meta-heuristic frameworks, approximation algorithms and carefully designed heuristics for specific instance classes to solve them efficiently. However, the resultant solutions can be very far from optimal, and the development of good algorithms often requires significant human effort. The goal of this PhD project is to augment the human ability to design good algorithms and data structures by using machine learning techniques to explore the search space efficiently.

Over the last two decades, a large number of neural network architectures have been proposed to deal with a range of NLP and image processing tasks. However, these architectures (e.g., CNN) heavily rely on temporal and/or spatial coherence in the input sequence. In recent years, researchers have attempted to adapt these frameworks for solving combinatorial optimization problems with limited success. The combinatorial optimization problems often have long-ranged and complex correlations in the input element sequence. Therefore, the traditional architectures do not generalize as well as they do for the NLP and image processing tasks. In this project, we plan to consider restricted domains of combinatorial optimization problems (e.g., geometric optimization problems) and design neural network architectures that can learn efficient features for problems from the domain and leverage those features to find good solutions for large instances of the problems.

Modal Title

Elham Mohammadzadeh Mianji

Student

Project Title

Proactive Machine Learning-based Methods for Intelligent and Secure Vehicular Networks

Project Description

Vehicular networks have a high potential in the creation of smarter cities, but also smarter roads. This potential relies on the on the wheels connectivity provided by vehicular networks that can also meet the always connected need of drivers and passengers as they are spending much of their daily time in their vehicles. Moreover, the vehicular network is considered to have a crucial role in the context of self-driving vehicles. Vehicular networks are based on “smart” vehicles that are able to communicate to each other and to the infrastructure via vehicle-to-vehicle (V2V) and vehicle-to-infrastructure (V2I) communications, known under the generic term of V2X communications, but also via other wireless communications technologies.

Security is an important concern in vehicular networks as malicious attacks can lead not only to material losses, or reputation loss, etc., but can actually lead to the loss of human lives. Vehicular networks have specific characteristics that pose challenges to security, such as high mobility, rapidly changing topology, large scale, heterogeneous environment, etc. The implementation of security mechanisms in vehicular networks usually introduce an overhead in communication. In particular, the proactive mechanisms lead to high signalling and communication cost. Hence, the state of the art is represented by reactive detection in order to balance the quality of service and security. Machine learning (ML) algorithms have been used successfully in network security in general, and vehicular networks in particular. In the latter case, they demonstrated their potential of dealing with the challenges imposed by the aforementioned unique characteristics of vehicular networks.

This project’s main research objectives are listed below, but the ultimate goal of the project is to propose proactive security methods for vehicular networks.

RO1 – comprehensive literature review in the area of ML-based security mechanisms for vehicular networks.

RO2 – propose proactive security methods for vehicular networks that are based on ML. Blockchain will be also investigated as it is showing promises in this area.

Modal Title

Fangyijie Wang

Student

Project Title

Automatic imaging biomarkers in foetal development using multi-task deep learning framework

Project Description

Ultrasonography is widely used in the field of obstetrics, which is a popular way of assessing the state of foetal development (e.g. gestational age (GA) estimation, foetal weight (FW) estimation) and safety of the pregnancy. The operator usually performs an array of measurements during ultrasound (US) scan sessions. The scans are safe and can be carried out at any stage of your pregnancy. Over the years, some researchers have demonstrated that deep learning algorithms are used to reduce operator dependent errors and improve the accuracy of foetal well-being assessment, but the use of AI is still in a stage of infancy.

 

Foetal biometric measurement is a standard examination during pregnancy used for the foetal growth monitoring and estimation of gestational age. The most important foetal measurements include the measurements of biparietal diameter (BPD), head circumference (HC), femur length (FL) and abdominal circumference (AC). To obtain these proper measurements, it requires the use of standardised planes. The operator needs substantial knowledge and experience to identify the standardised planes that are foetal abdomen (FASP), brain (FBSP) and femur (FFESP) standard planes. However, expert resources are scarce, especially in underdeveloped countries. We believe that deep learning algorithms can be a valuable tool to tackle these challenges. Biometry measurements are performed on ultrasound images that have standardised planes. After that, these biometric parameters can be used to evaluate foetus growth and estimate the following parameters: GA and FW.

 

The first phase of this research is using a literature review to understand the recent research results

related to foetal ultrasound deep learning applications. The second phase of this research is exploring multi-task neural networks for automatically classifying and segmenting foetal body parts in 2D ultrasound images. Foetal body parts include the foetal head, abdomen and femur. The third phase of this research is developing a novel deep learning framework for measuring the foetal body parts, BPD, HC, FL and AC. These foetal biometrics are used to assess foetal growth.

The results of this research may reduce the rates of misdiagnosis for foetus growth monitoring. Thus, it contributes to the improvement of the quality of medical services and ultimately benefits patients.

Modal Title

Harsh Singh

Student

Project Title

Exploiting Rich Long-term Memory Context in Spatiotemporal Vision Tasks without Annotations

Project Description

Videos are one of the richest and vastly produced data types. Due to its higher dimensionality and complexity, it is challenging to model spatiotemporal vision tasks with classical approaches. Recent research has focused on solving these tasks with deep learning techniques. Supervised learning has been the most successful version of these techniques, but it requires annotations to train the model. As video data is massively growing with time, it is impossible to create labels for different tasks to capture a wide variety of patterns. The research aims to solve this annotation scarcity in spatiotemporal vision tasks with self-supervised and unsupervised learning paradigms. In order to exploit memory context, recent research has used various internal memory modules. However, these internal memory modules do not exploit rich long-term memory elements. The research aims to solve this issue by introducing memory networks. Memory networks contain an external memory module synchronized with deep learning architecture. Unlike internal memory modules, they capture rich long-term past knowledge. On the other side, excellent qualitative results can also be obtained from these memory modules, proving their transparency and explainability. For achieving both aims, this proposal suggests testing various hypotheses obtained from memory networks and related literature on three different spatiotemporal vision tasks: (i) self-supervised video object segmentation (Self-VOS), (ii) video prediction, and (iii) unsupervised video anomaly detection.

Modal Title
Jiwei Zhang

Jiwei Zhang

Student

Project Title

Explainable Natural Language Processing for Legal Text Analysis

Project Description

As a result of developments in machine learning, particularly neural networks, a growing number of state-of-the-art technologies are employing deep learning to identify solutions to real-world problems. Due to the complexity of its real-world data, Natural Language Processing (NLP) is a domain in which deep learning techniques have become dominant, particularly for tasks dealing with long text documents.

The document-level classification task is a significant challenge in the research community of NLP because it has a wide range of practical applications, including legal text analysis, sentiment analysis and mapping labels for news articles. A key difficulty for document-level classification tasks is to understand the relations between sentences, which is not easily achievable by traditional approaches like regression models. To achieve document-level understanding, current approaches typically rely heavily on transformer-based neural network modules, such as BERT and its variants (e.g. DocBERT and RoBERTa), XLNet and GPT-3.

However, as the implementation of deep learning neural networks becomes more widespread, additional obstacles emerge. In the majority of instances, when a neural network is employed for downstream tasks, users are only privy to the predicted results but not reasons for those predictions. Neural networks are often referred to as “black boxes” since it is impossible to interpret the actual meaning of the weight matrix. In other words, people have difficulty interpreting the relationship between the inputs and the outputs. Even if the accuracy of the prediction outcomes may be the most important feature in some disciplines, the prediction method must be transparent, understandable, and interpretable.

For legal text classification, it is common that documents are long and domain-specific in terms of their vocabularies. For example, in the legal AI community, there is much research focusing on tasks like categorising legal cases based on legal opinions and legal regulation classification. Classification models generally have complex architectures and consist of several embedding modules and neural network modules, limiting interpretability. Therefore, it is rare for industries or legal departments to make use of these results directly in the real world because of a lack of understanding of why predictions have been made. Legal and business leaders are typically reluctant to rely on opaque models of this type in their decision-making processes.

To find a solution to this problem, the most vibrant area of research is eXplainable Artificial Intelligence (XAI). In the context of this project, we will investigate XAI approaches for long-text document classification tasks in the legal domain. Our research will include but not be limited to current reasoning approaches, state-of-the-art models for long-text classification tasks, interpretability of neural networks, machine learning on weakly-labelled data, and the application of these technologies within the legal domain. With this research project, we aim to maximise the advantages that deep learning approaches have brought, by achieving transparency and interpretability in long-text classification tasks in the legal domain, thus providing a more reliable base up which decisions can be made.

Modal Title

Mazhar Qureshi

Student

Project Title

XAI for hate speech monitoring on social media

Project Description

In this age of uninterrupted social media access, the extent of connectivity for an ordinary individual has reached unprecedented levels, allowing the spread of ideas to increase many folds and turning the world into a global village of public views and opinion. Social media’s low-cost and high-speed connectivity has made it a favourable avenue for alternative or alt views that may be underrepresented in mainstream media. Hate speech on social media has been an interesting research topic for many years under the broader umbrella of Internet Sciences. Most researchers define ‘hate speech’ as derogatory remarks towards an individual, race, religion, gender, or sexual orientation. However, what constitutes derogatory and what does not remain a much-debated topic.

Several techniques have been proposed in the artificial intelligence community to identify hate speech and disinformation. Several benchmark datasets contain data gathered from popular social media platforms, i.e., Twitter, Facebook and other platforms such as Gab and Whisper. This data is often biased due to the source of labelling or the models trained on these datasets’ failure to identify multiple hate-speech incidents. Conversely, many excerpts are falsely identified as hate speech as well. The lack of consistency in the definitions, labels and classification of hate speech, along with a lack of explainability behind classification, causes mistrust between social media networks and users.

The explainability of AI models refers to the degree to which an ML model is understandable to its stakeholders. Here, Explainable AI (XAI) aims to improve the user experience by increasing trust in the decision-making capabilities of a system. The introduction of explainability to purpose-built AI and ML solutions has been a sought-after concept for several years now. Around the world, policymakers have pushed for more explainable and transparent solutions to enable effective policy making using AI in critical systems. Similarly, an explainable model for hate speech detection can provide similar insights to policymakers for legislation on hate speech control on social media. Furthermore, explainable solutions also improve the public understanding of these frameworks and algorithms.

There is a need to address the lack of explainability in monitoring hate speech on social media. Establishing a certain level of trust between the system and its stakeholders is a genuine requirement. Similarly, explainability enables policymakers to develop better policies with an increased understanding of the system. The following research objectives define the scope of the project:

  • To analyse the current systems in place for hate speech detection, their implementation and a review of the relevant literature.
  • To develop a hate speech detection system that can classify hateful statements with minimum biases.
  • To explain hate speech by applying and adopting various effective XAI approaches that bring transparency into a policy violation.
Modal Title

Muhammad Mubashar Saeed

Student

Project Title

Machine learning based characterisation of nanoparticles synthesized using microfluidics

Project Description

The application of machine learning (ML) to microfluidics is currently gaining significant momentum within the research community particularly for applications in material science, nanomedicine and biologics. Microfluidics is the miniaturisation and automation of laboratory processes resulting in improved quality of data generated with reduced reagent and support costs. While most widely applied to diagnostic testing (e.g. pregnancy tests and COVID-19 testing). Microfluidics is also applied to high-throughput applications such as bioprocess monitoring, cell analysis, genetic screening, and drug discovery. These high-throughput applications involve large data sets of complex biological processes – thereby, making them ideally suited to control and optimisation through ML.

An emerging application of high-throughput microfluidics, particularly for drug discovery and in the development of drug delivery systems, is the use of microfluidics for nanoparticle synthesis. Furthermore, these same microfluidic platforms can also be used to investigate the interaction of nanoparticles with cells. These platforms can help optimise drug design and uptake, and therefore potentially improve patient outcomes for many chronic diseases (e.g., various types of cancer, HIV and diabetes).

The premise of the proposed research is based on ML application to microfluidic nanoparticle synthesis to optimise their desired properties (i.e., particle size and charge, polydispersity index, drug loading capacity and cell encapsulation). Then application of similar algorithms to optimise the nanoparticle interaction with cells for efficacious drug delivery and therapeutic response. In this project, the specific aim will be to implement different passive microfluidic techniques (e.g., hydrodynamic flow focusing, microvortices, chaotic advection and droplets generation) for efficient fluid mixing leading towards optimal nanoparticle synthesis. ML will be implemented through the development of at-line monitoring technology to measure the physical properties of the nanoparticles within the microfluidic channel. The coupling of this multimodal monitoring to ML is novel and can lead to innovative advances in the field of nanomedicine, biologics and organ-on-a-chip.

Modal Title

Naoise McNally

Student

Project Title

Auditing Algorithms: Investigations of the Facebook News Feed Recommender System

Project Description

In the past decade digital intermediaries such as Facebook, Google and Twitter, have assumed a pivotal role in the information space as central distribution channels for all forms of content, most notably news. The distribution and (in)visibility of information on these platforms are dictated by algorithmic recommender systems, which are often described as black boxes, in reference to the opaque nature of their decision-making systems. The information delivery outcomes of such systems are understood to have profound implications for the public discourse, yet oversight and transparency have been severely limited.

Efforts to understand the effects of such algorithmic systems have resulted in an emerging area of research: algorithmic auditing. Sandvig et al (2014) set out the initial understanding of algorithmic audits as a variety of methods used to uncover issues within the decision-making structure of an algorithm. As a nascent area of research, methods have yet to be standardised and encompass a wide variety of techniques investigating algorithmic systems by indirect means. Such audits by academics, journalists and activists in recent years have uncovered evidence of harmful algorithmic mechanics on various platforms including issues of racial bias, discrimination, misjudgement and misattribution.

The proposed research uses an empirical approach to investigate the effects of algorithmic governance of information on the Facebook platform, by auditing the Facebook News Feed recommender system. The research design includes using a parametrized timeline of known strategic changes to the Facebook News Feed recommender system, in conjunction with media content including a corpus from The Guardian newspaper (2011-2020), and utilising CrowdTangle to access Facebook engagement metrics for such content. The proposed method is to build a model based on the documented changes to Facebook algorithms, which is subsequently modeled with Cross-Correlation temporal analysis, Augmented Dickey–Fuller and Granger-Causality tests, and finally the Seasonal Hybrid ESD (S-H-ESD) algorithm. This study presents a proof-of-concept audit of the Facebook News Feed and forms the basis of an extended set of further investigations aimed at contributing to the understanding of algorithmic governance on digital intermediaries.

Modal Title

Nighat Bibi

Student

Project Title

Explainable Arficial Intelligence (XAI) in Healthcare

Project Description

The brain is the body’s command centre; it controls the function of each organ. The effects of any disruption in the brain can be disastrous. Therefore, it is essential to find brain illnesses early before they deteriorate. Brain tumour, Alzheimer’s disease, Autism, and other common brain conditions must be recognized in the early stages; otherwise, the outcome may be worse.

Brain tumour occurs because of the abnormal development of cells in the brain. It is one of the significant reasons for death in adults around the globe. Millions of deaths can be prevented through the early detection of brain tumours. MRI images are considered helpful for detecting and localising tumours.

Alzheimer’s disease is a degenerative neurological condition that causes the brain to atrophy, which causes the brain to shrink and the brain cells to die. It affects people between the ages of 30 to middle 60. Alzheimer’s disease affects 5.8 million people in the United States who are 65 years or older. It is a typical dementia cause. Sadly, Alzheimer’s is incurable and can cause death and a severe loss of brain function. Therefore, it must be detected early and treated.

Autism spectrum disorder (ASD) is a neurological disorder that impacts how people connect with others, communicate, learn, and conduct. It first manifests in early childhood, evolves throughout life, and needs to be caught early to speed up therapy and recovery. In addition, medical brain imaging techniques may be used to identify these impairments.

There are different biomedical image techniques. However, MRI images provide clear images of a brain that can help an accurate diagnosis of brain diseases.

Many AI-based approaches already exist for diagnosing brain diseases; however, the black-box approaches are not considered more reliable in the healthcare field, so the explainability of AI-based models is crucial in disease diagnosis. Explainable Artificial Intelligence supports researchers in justifying their model with transparent results that lead to trustworthiness for clinicians, doctors, and patients.

Objectives:

We aim to provide explainability of the diagnosis of brain diseases, i.e., Brain tumours, Alzheimer’s Disease, and Autism, from MRI images. The fundamental reasons behind this research are:

  • Provide accurate, fast, and early detection of brain diseases
  • Provide a transparent/trustworthy/explainable diagnosis of brain diseases (why and how our model predicts these results)
  • Detect more than one type of brain disease from MRI images
  • Proof that AI-based models are trustworthy for the diagnosis of diseases from MRI images

Approach

In this research, machine learning and deep learning models will be employed to diagnose brain diseases (Brain tumours, Alzheimer’s disease, and Autism) with high accuracy from MRI images and XAI methods (like SHAP, LIME, and LRP) will be used to provide transparency of the models and reason behind their decision (output).  

Modal Title

Nils Hohing

Student

Project Title

Grounded Language Understanding

Project Description

Natural language understanding systems rely on word statistics performed by language models. These approaches capture an approximation of language semantics, but they exhibit many known failure cases like poor understanding of causal relationships, being sensitive to rephrasing of sentences and producing syntactically convincing, but semantically questionable outputs.

Many of these issues can be attributed to language models’ lack of grounding in reality. Humans know which concepts from our world the words of a text correspond to, e.g. for the word “tree” what a tree looks like, which sounds it produces and which tactile impressions are associated with it. This knowledge gives us an edge over current language understanding systems in reasoning which is implicitly required for all language understanding tasks.

Existing research has contributed a variety of benchmarks to measure the alignment between different modalities like vision, language and audio. The best way to test for alignment are retrieval benchmarks like Winoground, where the model is tasked to retrieve items like images from a big database that best match a key, e.g. a given text. Image or text generation benchmarks in contrast have unreliable automatic evaluation because defining a sensible distance metric between a ground truth image or text and a generated one is very hard.

Learning the alignment currently works very well for higher level concepts, for example understanding the visual differences between a wolf and a bear, but it fails in the details. For example simple spatial understanding like discerning between left and right surprisingly often does not work. Also unusual compositions are rarely understood well. For the prompt “a cup on a spoon” DALLE-2, the image generation model, generates only spoons in cups. This reveals serious deficits in the model’s language understanding.

This project aims to overcome failure cases of those existing solutions by improving models that understand the relationship between words and images (possibly also videos) measured by existing benchmarks. Additionally, new benchmarks to measure the performance in those areas more precisely will be created. At last the point is to demonstrate that these image-text multimodal models can outperform language models in purely textual domains (when there is no visual information available at inference time).

For the first step there could be three sources of the aforementioned problems with image-text models: the model architecture, the data and the learning strategy.

-The initial experiments have shown that the image and text processing architectures are capable of learning basic physical relations like “left” and “right”

-Since the datasets used in this domain contain millions up to billions of image-text pairs, a lack of data also seems unlikely.

-Therefore, either the quality of the data or the learning strategies must be the problem.

To start, the goal of this project is therefore to examine the data quality for the specific purposes and to improve image-text alignment via novel curriculum learning strategies.

The main challenges will be working with very big datasets and doing meaningful evaluation.

Modal Title

Reyhaneh Kheirbakhsh

Student

Project Title

Semantic Deep-Learning Approach for Medical Image Analytics

Project Description

Gliomas are among the most aggressive primary tumours that occur in the brain and spinal cord. A glioma can affect your brain function and be life-threatening depending on its location and rate of growth. Gliomas are classified according to the type of glial cell involved in the tumour, as well as the tumour’s genetic features, which can help predict how the tumour will behave over time and the treatments most likely to work. Therefore, identifying the type of glioma will help determine appropriate treatment and prognosis. In this project, we propose to use data analytics to identify the type of glioma tumour using an elaborative mining process. The process requires data collection, data pre-processing and segmentation, application of deep learning and interpretation of the results. The innovative part of this project is in its ability to incorporate semantic elements in the learning process so that the results are reliable with high accuracy. This study will be extended to other types of medical images to identify other types of tumours and injuries.

Modal Title

Sukanya Mandal

Student

Project Title

Privacy Preserving Federated Learning based Cognitive Digital Twins for Smart Cities

Project Description

Recent advancements in AIoT (Artificial Intelligence + Internet of Things) techniques have made possible continuous advancements of various smart city applications in various sectors. Cognitive Digital Twin (CDT) – a Knowledge Graph and AI based virtual replica of the physical world, has been adopted by various industries especially the manufacturing sector but has been significantly slow adoption for smart city. The major reason being – lack of trust and privacy concerns towards sharing sensitive data. Privacy Preserving Federated Learning (PPFL), could be integrated along with CDT to ensure privacy preservation and trustworthiness. This research proposes a framework for integrating PPFL and CDT technologies to address various real life smart city scenarios as well as enable feasibility for smart city governance.

A CDT is an extended or augmented version of DT with cognitive capabilities. It contains at least the three basic elements of DT – the physical entity (systems, subsystems, components etc.), digital (or virtual) representation or shadows and the connections between the virtual and physical spaces. The main difference is that CDT usually contains multiple DT models with unified semantics topology definitions. In addition, a CDT should – have cognition capabilities, have a digital version of the entire lifecycle of a system, be autonomous and be able to continuously evolve with the physical system across the entire lifecycle. The core idea of Federated Learning (FL) is to train machine learning models on datasets that are distributed across different devices or parties, which can preserve the local data privacy. Privacy Preserving Federated Learning (PPFL) focuses on the privacy preserving mechanism of FL by employing privacy preserving techniques such as Homomorphic Encryption (HE), Secure Multi-party Computation (SMC) and Differential Privacy (DP).

The motivation behind coupling PPFL and CDT are quite a few. Even though CDT has its own benefits, it lacks maintaining data privacy as data has to be transferred to the CDT from various sensors and physical entities. In PPFL, the privacy is preserved at the end user as only weights and parameters updates are transmitted from the edge devices and not the data. Hence raw sensitive information stays with the device and is not communicated over the network. The communications between smart city DTs as well as between DTs and edge devices is privacy-preserved. Last but not least, data quality and integrity is improved as pre-processing is done at the edge environment.

Research Objectives and Directions:

  • Create Ontology and Knowledge Engineering models for smart cities (serving as the data layer).
  • Research on creating a GNN based framework (or other enabling technology) to form the cognition capability of  the CDT framework basis graph data models catering to smart cities.
  • Research on a framework creation of Privacy Preserving FL in general and specialise it for adoption of CDT.
Modal Title

Svetoslav Nizhnichenkov

Student

Project Title

Human-in-the loop model training with explanations and feedback

Project Description

This PhD will work on human-compatible AI systems with a focus on trust and user-induced feedback loops. Recent advances in explainability and interactive AI have opened up avenues for users to influence AI systems, however, this presents many challenges such as how to quantify knock on effects, how to seek feedback from users, how to efficiently incorporate user feedback, and how to facilitate AI systems to negotiate a common objective and understanding between the system, the domain expert user and the AI system designer. This PhD will explore research challenges in this space with a view on solving real world problems that AI system designers face.

Modal Title

Van Hoang

Student

Project Title

Style and Personalisation in Situated Interaction

Project Description

Since the appearance of Siri from Apple, dialogue systems have become more and more prominent in our lives. Recently, there has been an increasing interest in the Natural Language Processing (NLP) community to design adaptive systems. Initial research has shown that stylized and personalized conversations, tailored to users’ needs and preferences, would help strengthen the connections between dialog systems and human users. As a result, personalized content improves user engagement in conversations, increases communication effectiveness, and develops trust in the systems.

For the selection of user-centred content, psychologically motivated concepts such as emotions and personality have been investigated and incorporated into the development of human-like conversational dialogue systems. In contrast to short-lived emotions and affective states, personality traits are more stable and endurable over time. Therefore, personality is better suited to model long-term user preferences while emotions are for short- and mid-term preferences. Injecting human traits into a system should start with understanding real human interactions. However, there is seemingly a lack of insights from other disciplines in popular research literature. The definitions of “emotions” and “personality” are often data-driven, and so are the responses of the systems to the users. In the PERSONA-CHAT dataset, a persona, or personality, is a list of five random characteristics.

There are three key challenges in delivering stylized and personalized content to users by emotion- and personality-aware dialogue systems.

The first one involves the automatic detection of the user’s affective states and personality traits to build their models. Which emotion and personality inventories should be selected? And what could be used as feature cues for the detection (e.g. texts, speech, body language)? Secondly, with the user data from the previous step, the generated responses should be personalized and stylized to user preferences. Furthermore, the personality of the systems should be consistent throughout the conversations. The last challenge is about the ethical aspects of the dialogue systems. For example, given the users’ distressing emotional states, how should the systems respond? More importantly, when should they try to change user behaviours, and when not? 

Using established theories from both psychology and linguistics, and latest model architectures from NLP, the project aims to partly address the second and third challenges. In the Style Transfer task, the GAN architecture has been utilised extensively for the conversion of texts from one style to another according to users preferences. This method lies in the assumption that style and content can be separated completely . However, recent work has proven that such clear separation is not easily attainable, if not impossible, depending greatly on the domains. These findings have motivated us to examine other frameworks for a deeper understanding of their own strengths and weaknesses. Taking a data-centric approach to the challenges, we work to develop flexible ML applications in NLP that can deliver these goals.