Introduction
One key difference between human and artificial intelligence is the number of training examples needed to generate knowledge. Whereas humans can learn to recognize new objects with only a few examples, most machine learning tasks require hundreds of examples for the same task. In fact, increasing the dataset size is often a key step in improving the performance of a machine learning model. ImageNet [
1], the most famous dataset in computer vision, now consists of over 14 million training examples. The state-of-the-art models in computer vision are often trained on large datasets such as ImageNet and may not transfer well to smaller datasets of different tasks. Getting large datasets may not always be a feasible approach though, especially in the medical domain.
Gathering large datasets is one of the key challenges of medical deep learning applications. Keeping a patient’s medical information safe is critical and there are laws protecting it in most countries. This makes it more difficult to get the data and leads to the medical datasets being much smaller compared to traditional computer vision datasets. Additionally, deep neural networks themselves offer another privacy threat. It has been shown that training examples of fully trained networks can be recovered with a model inversion attack [
2]. This makes it more difficult to publish medical deep learning applications as the patient’s privacy can not be guaranteed. These two reasons give a big incentive to find ways to train neural networks with smaller datasets or even just one patient’s data.
There have been several models proposed to challenge the task of reducing the number of training examples. One-shot learning is a method of learning a class from only one labeled example [
3]. Siamese neural networks are able to determine if two images show the same person, even if they have never seen images of that person before [
4]. They have also been used in medicine to distinguish between chronic obstructive pulmonary disease and asthma [
5]. Whereas new classes can be learned from as little as one example, one-shot learning still requires thousands of training examples of other classes beforehand. Furthermore, anomaly detection can be used to detect classes of rare occurrence. This is a technique used to recognize items which do not lie in the usual data distribution and makes use of unsupervised learning in most cases [
6]. Anomaly detection usually makes use of learning the data distribution in a healthy population and identifying the anomalies, i.e. a disease, of a new class. Another method to handle small datasets is transfer learning, where networks trained on large datasets are used as a starting point to train on training examples of new classes. Transfer learning makes use of the fact that features learned on the large dataset can be reapplied to new data.
In this paper, we introduce personalized neural networks, which use only one patient’s data for training. Our proposed method only needs two MRIs from the same patient and no additional pretraining. This also results in a privacy-safe processing of the data, because the data “stays” within the same patient. Our model is based on generative adversarial networks (GANs) [
7]. GANs have gained in popularity in recent years in the medical AI community. Originally used for image synthesis, there have been applications to generate medical images [
8,
9]. Other studies focus on classification or segmentation tasks [
10,
11]. We apply the personalized neural networks on subjects with brain tumors.
Brain tumors belong to the most devastating diagnoses, in particular for a confirmed glioblastoma multiforme (GBM) [
12]. Despite massive research efforts and advancements in other cancer types, like breast cancer [
13] or prostate cancer [
14], the life expectancy of a confirmed GBM with treatment, including chemotherapy, radiotherapy and surgery, is still only around one year [
15]. Nevertheless, disease progression and treatment decisions are strongly dependent on maximum tumor diameter and tumor volume, as well as the corresponding morphological changes during a treatment period. The imaging method of choice here is magnetic resonance imaging (MRI). However, MRI does not provide any semantic information for brain structures or the brain tumor per se. This has to be done manually, semi-manually or automatically, in a post-processing step, commonly referred to as a
segmentation. Manually performed, however, a segmentation is very time-consuming and operator-dependent, especially when performed in a three-dimensional image volume [
16], which needs slice-by-slice contouring. Hence, an automatic (algorithmic) segmentation is desired, especially when large quantities of data volumes have to be processed. Even if it is still considered an unsolved problem, there has been steady progress from year to year; and data-driven approaches, like deep neural networks, currently provide the best (fully automatic) results. However, a segmentation with a data-driven approach, like deep learning [
17], comes with several burdens: Firstly, the algorithm generally needs massive annotated training data. Additionally, for inter-patient disease monitoring, several segmentations have to be performed, and in addition, these scans have to be registered to each other (which also adds uncertainty to the overall procedure, especially when deformed soft-tissue comes into play [
18]). In this regard, we want to tackle these problems with a personalized neural network that needs just the patient’s data, no annotations and no extra registration step.
We apply the personalized networks to longitudinal datasets of glioblastoma. To the best of our knowledge, this is the first study using this little training data to train a deep neural network in the medical domain. The method addresses the issues of gathering big datasets in medicine and producing a privacy-safe network. The approach is considered as unsupervised learning as no data annotation is necessary. Using a Wasserstein GAN, the model creates a map showing the differences between images from two timepoints. We evaluate the model with an receiver operating curve (ROC) analysis as well as a modified RANO criteria on two different datasets of longitudinal MRI images of patients with glioblastoma.
Discussion
In this contribution, we propose “A net for everyone”, a personalized neural network that is trained with longitudinal data from a single patient. We designed and implemented a Wasserstein-GAN-based approach that works with only two scans from the same patient without any extra training data in an unsupervised fashion. That means, our method does not need any small or large quantities of datasets, and also does not need any manual or semi-manual annotations for training.
Alongside a qualitative evaluation, we show that the model achieves a high AUC in an ROC analysis, when compared to a state-of-the-art deep learning model. It also shows that the model’s performance for tumor growth and tumor reduction is very similar. The accuracy for the local dataset was significantly larger than for the public dataset. This can be explained by the difference in quality, as the public data was older and had a lower resolution, especially in the third dimension. Additionally, there were artifacts in some of the images, like parts of the brain were cut off. We implemented a modified RANO criteria, resulting in a combined accuracy of 66%. The generated heatmaps can aid in the diagnostic process to quickly find the key regions of interest.
It should be noted that the performance of deep learning models usually scales with the size of the dataset [
35]. Therefore, this approach has an inherent disadvantage compared to classical supervised learning models with big datasets. However, using only the data of one patient comes with some advantages. First, our method is a privacy-safe approach. Medical records and medical image data are very sensitive and our approach stays within the same patient for the algorithmic training and execution. Second, getting large datasets in medical imaging has proven to be a challenging task due to these privacy concerns, and our method does not rely on this.
Furthermore, no registration is necessary for the training of our approach, which is a mandatory and crucial step in most approaches [
36]. There are different methods for image registration, with some being completely automatic and others needing some manual input [
37]. While these registration methods can be accurate for scenarios, like rigid registrations, especially deformable registrations are still challenging and there are problems with outliers [
38]. These include post-surgery scans or patients with a different anatomy due to a large tumor. Both could lead to registration artifacts, which would compromise the further training. Our model does not need a separate registration step, avoiding these potential sources of errors.
The model does not explicitly learn to recognize changes in the tumor, but learns to recognize any changes between two images. However, since the contrast enhancing regions of the tumor are typically amongst the most intense regions in a T1ce scan, changes in these regions are particularly visible in the created maps, highlighting changes in tumor enhancement patterns. However, the proposed approach comes with two disadvantages that can be addressed in future research. First, any structural change in the brain not lying in the tumor will be recognized by the model. For example, a midline shift caused by tumor growth will cause changes in healthy regions of the brain and might be interpreted as growth or reduction of contrast enhancing tumor. This can also be interpreted as an advantage to point out all changes to the reader. Second, the model is prone to noise at the edge of the brain and next to the ventricles. The ventricles differ between two scans depending on the current cerebral spinal fluid volume. At the edge of the brain, the two scans also differ slightly due to the skull stripping. Another reason is the variance in size of the dural venous sinuses. To account for the noise at the edge of the brain, we disregarded the outer pixels in the calculation of the modified RANO criteria. This is obviously a concern for tumors located in the cortex of the brain as it might cut out regions of the tumor. However, glioblastoma are typically located in the centrum semiovale, thus in most cases this should not be a problem [
39].
It should be noted that the ground truth from this work was not created by medical experts but by a neural network. However, the network used achieved a Dice Score for the enhancing tumor of 82% [
33]. This lies within the range of the inter-rater variability of human raters of 74–85% [
40], suggesting that medical experts would not change the ground truth significantly.
However, despite the above-mentioned limitations, this study is a proof of concept that personalized neural networks can serve as a privacy-safe method to analyze longitudinal imaging data of a single patient in an unsupervised fashion. It has been shown that tumor growth tends to get underestimated on average and overestimated for very small tumors in brain tumor measurements in the current RANO criteria [
41,
42]. Therefore, having an efficient method for measuring the 3D tumor volume is necessary for treatment monitoring and surgical planning [
43,
44]. Lastly, the produced heatmaps can be a big help in the diagnosis of the MRI images, as they lead the reader directly to the key regions of changes.
Summarized, we proposed a deep learning architecture to create personalized neural networks. This study serves as a proof of concept to show that training data from just one patient can be used to monitor tumor change in longitudinal MRI scans. Areas of future work include the application to other pathologies, such as aortic aneurysms and aortic dissections [
45], where disease monitoring over several image acquisitions plays an important role.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.