(2019-31.12.2023)
Artificial intelligence (AI) is making impressive progress in all sectors, including healthcare. Indeed, by means of AI technologies, we can improve the healthcare system and gain insights into diseases we would not be able to gain without it. In 2019, the Department of Economy, Science and Innovation provided impulse funding for cutting-edge AI research: the Flanders AI Research programme.
The research program is organized around four grand challenges and within this program, Liesbet M. Peeters acts as the use case lead of the Use Case multiple sclerosis (MS), which is part of Grand Challenge 1 (AI-Driven Data Science: Making Data Science Hybrid, Automated, Trusted and Actionable), Work Package 7 (use cases in Health). We aim to speed-up the identification of the right treatment for the right patient at the right time by improving the data management of data that is already collected and applying AI techniques on these datasets.
To do so, we collaborate with the other Flemish universities: University of Leuven (KU Leuven, including campus Kortrijk (KULAK)), Ghent (UGent) and Antwerp (UAntwerp), more specifically following research groups are involved:
PI’s involved in our collaborations: Alexander Bertrand (KUL), Luc De Raedt (KUL), Maarten De Vos (KUL), Yves Moreau (KUL), Johan Suykens (KUL), Celine Vens (KULAK), Tom Dhaene (UGent), Willem Waegeman (UGent), Jan Sijbers (UAntwerpen).
Within the Use Case MS, we aim to meet three clinical challenges:
To meet the three clinical challenges, we work on 4 different Proof of Concepts (POC):
Summary:
In 2020, we used dockerized Jupyter notebooks to run the federated pipeline of the GDSI (MSDA infrastructure v1.0). However, since then we have been working on improving the MSDA federated architecture (MSDA infrastructure v2.1) to incorporate some of the feedback, comments and suggestions we have received. The main differences between the MSDA infrastructure v1.0 and v2.1 are the following:
More details about the MSDA infrastructure v1.0 and v2.1 are provided in this link.
Next to this, this video provides you with a demo of the MSDA infrastructure v2.1.
A paper about the federated pipeline used in the GDSI was submitted to a scientific journal and is currently under review.
Summary:
Our primary objective is to evaluate the potential of implementing Federated Learning in the intricate multiple sclerosis ecosystem and identify the most effective federated methods. This endeavor holds immense significance as it has the potential to pave the way for groundbreaking advancements and foster unprecedented innovation in this field. This project aims to comprehensively assess the impact of federated learning on predicting disability progression in people with MS and compare its effectiveness to the well-established MS Benchmark. The first step involves partitioning the centralized MSBase data into virtual clients and constructing various scenarios to evaluate the effect of different designs on training performance. Subsequently, different federated strategies will be employed to determine their effectiveness using diverse approaches. Three key metrics, including predictive performance, time to convergence, and computation cost, will be assessed and compared to the centralized scenario, ensuring a scientifically rigorous analysis.
Deliverables:
Partners involved: UHasselt, Noorderhart, KU Leuven
Summary:
The study of longitudinal patient data has received much attention. The increased availability of such data reinforced this trend, allowing data-hungry machine learning techniques to be applied with a focus on going further towards patient specific precision medicine. However, temporal patient data present a number of challenges that are not easily addressed. The main difficulties are missing values (some measurements are not available for every patient), the non-constant sampling (time between hospital registrations can be very sporadic) and the joint nature of the data (events and longitudinal data are both present). We propose to address these issues with a new modeling technique relying on Bayesian dictionary learning of latent processes. This methodology posits that the observations made about each patient are generated by an underlying hidden disease process, accounting for the trajectories variability among patients.
Deliverables:
Partners involved: UHasselt, Noorderhart, KU Leuven
Summary:
Delays in identifying treatment failure, and in selecting a more optimal DMT from the next line, reduce the quality of life of people with MS.
Doctors need help to find the right treatment and to detect treatment failure (="non-responders") more quickly. We believe there is an urgent need for accurate and high-performance algorithms that can support decisions around progression and treatment of the disease. Moreover, these algorithms should be developed using "observational datasets". Observational datasets bring together data from a heterogeneous population (in contrast to clinical trials).
However, although there is great potential in using observational datasets, this is also accompanied by several challenges because of the many imperfections inherent in this type of data. Therefore, there is a need for new algorithms that can better deal with these imperfections. In the context of another project ("Supporting treatment decisions for multiple sclerosis in daily practice with high-performance systems”), we achieved very promising results. Indeed, we managed to greatly increase the accuracy of predictions of progression by introducing machine learning methods. Nevertheless, there is still room for improvement in this work. In this project, we therefore focus on the following 2 problems: 1° machine learning methods today can only be developed and implemented by technical experts and 2° when the datasets are not very large, the accuracy is limited.
Deliverables:
Partners involved: UHasselt, Noorderhart, UGent
Summary:
In this project, we aim to support research into computer-assisted methods for patient follow-up and evaluation of the effectiveness of therapies by looking at a mainstay of computer-assisted methods: data. To obtain reliable results, these methods need to be provided with quality data. Our focus will be on data resulting from a commonly used clinical test in MS patients: evoked potential (EP) data. Existing methods around quality assessment and improvement for this data modality are very limited and our aim is to improve on this. Our hypothesis is that quality assessment and improvement of EP data leads to a higher level of reliability in current computer-assisted methods. First, we investigate methods for automatic quality assessment and evaluation and, in a second step, we look at methods for quality improvement in EP data identified in the first step as not being of sufficient quality for further processing.
Deliverables:
Partners involved: UHasselt, Noorderhart, KU Leuven
Summary:
Whether or not a treatment is the right treatment for a particular person is very difficult to determine (treatment-failure). One of the important reasons why it is so difficult to estimate the effectiveness of treatment is because today we do not have the right outcome measures to measure treatment effectiveness. Today, the outcome measure "EDSS" (expanded disability severity score) is usually used. However, the EDSS outcome measure is not sensitive and therefore changes are often noticed very late (e.g. only after 2 years or even longer). Previous research by our research group has shown that evoked potential (EP) data can be supportive in predicting progression.
With this research, we aim to further improve the models by developing and implementing new algorithms. In doing so, we are building on an already approved protocol (Modelling and predicting the progression of Multiple Sclerosis using retrospective data: a pilot study). We wish to improve these models using artificial neural networks (ANN), which could find certain complex relationships present in the data and are harder (or impossible) to find with traditional methods. In addition, we could design the models so that they are easily and intuitively explainable. This would provide great added value for the physician in charge of the patient.
Deliverables:
Partners involved: UHasselt, Noorderhart, UGent
Summary:
In this project, we aim to develop new advanced artificial intelligence (AI) techniques and showcase the relevance of these techniques in improving the performance of decision-support systems (DSS) for disease progression for people with multiple sclerosis (PwMS).
We believe large-scale and advanced modeling of real-world data (RWD) is needed to support the development of high-performance decision-support systems for multiple sclerosis.
Due to the large number of patients available in the existing registries, several works have used machine learning for various tasks). First results of the project regarding patient trajectories (Edward De Brouwer et al.) are very promising. Using the MSBase Global Dataset, we showed that with machine learning methods suited for patient trajectories modeling, we are able to predict disability progression of patients in a 2 years horizon with an AUC of over 85%, which represents more than 15% increase compared to baselines methods using static clinical features.
It is clear that the MSBase Global Dataset is the perfect data source to develop and showcase complex AI algorithms and that introducing more complex AI in the MS RWD increases the performance of tasks. Next to this, the MSBase consortium gives us the opportunity to work with the best clinical experts in the field, thereby greatly increasing the changes of this interdisciplinary research project.
in this project, we focus on overcoming following challenges:
Deliverables:
Partners involved: UHasselt, Noorderhart, UGent, KU Leuven (including KULAK)
Summary:
We want to develop new and innovative methods using ambitious and advanced methods to evaluate the effectiveness of short- and long-term therapy. In this project, we focus on investigating the relevance of two clinical tests routinely collected at the Noorderhart Rehabilitation and MS Center in Pelt: evoked potentials time series (EPTS) and magnetic resonance imaging (MRI).
Our hypothesis is that short-term differences in radiomics and/or evoked potentials time series features can predict long-term treatment efficacy.
First, we create two pseudo-anonymised MRI datasets linked to clinical data using retrospective data from two different hospitals: Rehabilitation and MS Centre Overpelt (RMSC) and Zuyderland Medical Centre Sittard (ZMCS). Retrospective MRI images are selected for patients with extensive follow-up data. The relevance of radiomics and EPTS is investigated cross-sectionally, longitudinally, as well as in a treatment effectiveness study.
Deliverables:
Partners involved: UHasselt, Noorderhart, Zuyderland Medisch centrum Sittard, UMaastricht, KU Leuven, UGent
Summary:
In the Radiomics/Epomics project, we are currently working with a consortium to identify new biomarkers in MRI and evoked potentials that should improve predictions around progression and effectiveness of treatments. The magnetic resonance (MR) imaging (MRIs) used as part of this project were taken during "standard clinical practice". That is, the MRIs were taken in the context of care and thus using few standardised protocols and other scanners. The features used to perform the task will vary depending on the scanner that was used. Specifically, images of a person obtained using the same scanning protocol will still have different features. The site-specific variation present in the features is undesirable in machine learning algorithms.
In this project, we investigate machine learning-based methods to reduce this unwanted variation (also called harmonisation). The removal of the unwanted site-specific variation by the methods found should lead to more accurate prediction of disease progression.
Deliverables:
Partners involved: UHasselt, Noorderhart, UAntwerpen, UMaastricht, Icometrix
Summary:
In Noorderhart, EPTS (evoked potential time series) are measured in people with MS using EPTS devices. However, the raw data of those measurements are locked in these devices. KU Leuven's Declarative Languages and Artificial Intelligence (DTAI) research group has written an algorithm that can read raw data stored in a binary format. This software uses pattern recognition and tries to identify both hierarchical structure and primitive data types without supervision. Since variable names are lost due to encoding, this data can be annotated in collaboration with an expert to link certain values back to specific measurements. To evaluate the algorithm, researchers at the DTAI will use a large amount of the raw data files (binary, serialized data dumps). More specifically, retrospective data from Noorderhart and UHasselt (EPOMICS, raw data from EPTS measurements up to 2017).
Deliverable:
Having an algorithm capable of unlocking EPTS data.
Partners involved: UHasselt, Noorderhart, KU Leuven
Summary:
The aim of this research project is to give a proof-of-concept of an approach that speeds up the integration of real-world health data. More specifically, we want to showcase the possibility of leveraging (semi-)automated solutions for data integration tasks developed by the data science community to the medical informatics domain. Given that such applications are scarce in the medical informatics community, we will need to explore possible solutions first, and apply those to real-world health data taking the sensitive nature of real-world health data into account.
From the wide range of data integration tasks, we choose schema mapping and data harmonisation to apply (semi-)automated solutions for health data integration. We need to consider two work packages in this research project:
Deliverables:
Partners involved: UHasselt, Noorderhart