Abstract

Deep learning has shown potential in significantly improving performance for undersampled magnetic resonance (MR) image reconstruction. However, one challenge for the application of deep learning to clinical scenarios is the requirement of large, high-quality patient-based datasets for network training. In this paper, we propose a novel deep learning-based method for undersampled MR image reconstruction that does not require pre-training procedure and pre-training datasets. The proposed reference-driven method using wavelet sparsity-constrained deep image prior (RWS-DIP) is based on the DIP framework and thereby reduces the dependence on datasets. Moreover, RWS-DIP explores and introduces structure and sparsity priors into network learning to improve the efficiency of learning. By employing a high-resolution reference image as the network input, RWS-DIP incorporates structural information into network. RWS-DIP also uses the wavelet sparsity to further enrich the implicit regularization of traditional DIP by formulating the training of network parameters as a constrained optimization problem, which is solved using the alternating direction method of multipliers (ADMM) algorithm. Experiments on in vivo MR scans have demonstrated that the RWS-DIP method can reconstruct MR images more accurately and preserve features and textures from undersampled -space measurements.

1. Introduction

Magnetic resonance imaging (MRI) is a noninvasive imaging technology that can provide structural, functional, and anatomical information for clinical diagnosis. However, its slow imaging speed may result in motion artifacts and image quality degradation, as well as lead to patient discomfort. To accelerate MRI scans, researchers are seeking methods to increase imaging speed by reducing the amount of acquired -space data without degrading the image reconstruction quality.

Accelerated MR image reconstruction from undersampled -space measurements is, in essence, a highly underdetermined inverse problem. Reconstruction methods based on signal processing have evolved rapidly over the past decades and can now explore and utilize the prior information about the desired MR image to achieve the reconstruction by using regularization methods under the premise of ensuring the uniqueness and stability of the solution. Sparsity is a commonly used prior information with the emerging popularity of Compressed Sensing (CS) theory [13], including fixed sparse transform (e.g., wavelet or/and gradient) [46] and more flexible adaptive sparse representation (e.g., data-driven tight frame [7] and dictionary learning [810]). High-resolution reference images obtained in advance in practical application scenarios can also provide prior information. They can provide structural similarity for the target MR images and obtain more sparse difference images [1113]. In addition, the structured priors, such as image support information [1416] and structural sparsity (e.g., group sparsity, block sparsity, and tree sparsity) [15, 17, 18], can be introduced into a reconstruction model based on the union-of-subspace sampling theory [19], which has been verified to be efficient in improving reconstruction accuracy.

In recent years, deep learning has received a great deal of attention in the field of medical imaging, especially for segmentation, denoising, classification, and acceleration of MRI tasks [20]. MRI approaches based on deep learning can be either data-driven or model-driven [21, 22]. Data-driven approaches are aimed at learning the mapping from undersampled -space/images to fully sampled -space/images [2328]. Model-driven approaches start from MR image reconstruction models and import the procedure of iterative reconstruction algorithms into networks [2932]. To ensure the quality of reconstruction performance, both approaches require pre-training processes with the aid of large, high-quality patient-based datasets. However, this is a challenge in clinical applications because it is difficult to obtain sufficient amounts of patient-based MR datasets due to patient privacy concerns.

Recently, Ulyanov et al. proposed a Deep Image Prior (DIP) framework [33], which demonstrates that convolutional neural networks (CNNs) have the inherent ability to regularize various ill-posed inverse problems without pretraining [34]. DIP can achieve satisfactory results by applying untrained networks with random noise as the network input. DIP has been used for denoising, inpainting, super-resolution reconstruction [3538], CS recovery [39], and medical imaging, such as PET image reconstruction [34], CT reconstruction [40], and dynamic MRI [41].

In this paper, we propose a novel deep learning-based Reference-driven method using Wavelet Sparsity-constrained DIP (RWS-DIP) for CS-based undersampled MR image reconstruction, which can achieve improved performance without any pre-training procedures. Our proposed RWS-DIP method incorporates structure and sparsity priors into a DIP framework and utilizes the priors to further improve the efficiency of learning. It not only builds a bridge between the constrained reconstruction method and deep learning, but also largely reduces the dependence on patient-based datasets and contributes to the expansion of clinical applications. Experimental results have shown that the proposed RWS-DIP method can obtain more accurate reconstruction than traditional DIP, particularly in preserving image textures and features. The main contributions of this paper can be summarized as follows: (1)The proposed RWS-DIP method utilizes both structure and sparsity priors of MR images. The former is introduced by using a high-resolution reference image obtained in advance as the input of CNN, whose structure is similar to target MR images and thereby incorporates structural information into network. The latter is used by regularizing the norm of coefficients in a wavelet domain to further enrich the implicit regularization of traditional DIP, which is enforced by the fixed network structure. These priors improve the efficiency and effectiveness of deep learning and contribute to the improvement in reconstruction performance(2)The proposed RWS-DIP is a novel deep learning-based MR image reconstruction method inspired by traditional DIP and does not require any pre-training. This advantage renders the training datasets unnecessary, which has significance in clinical applications

The remainder of this paper is organized as follows. Section 2 presents details on the proposed RWS-DIP method, as well as a review of traditional DIP. Section 3 includes experimental results from in vivo MR scans and also includes details about data acquisition, undersampled schemes, and the experimental setup. Section 4 provides a summary of the paper’s main points and its results.

2. Methodology

2.1. Traditional DIP for Undersampled MR Image Reconstruction

Applying traditional DIP to undersampled MR image reconstruction, the object function is where is the undersampled -space measurements of the desired MR image , denotes an undersampled Fourier transform operator, and is the norm. is an untrained deep CNN parameterized by , with the random noise as input.

The desired MR image can then be reconstructed by

The training of the network parameters is performed by solving the optimization problem in Equation (1) iteratively, which is guided by the attempt to best match the network output to the measurements in -space. In DIP, no pre-training procedure is needed and the network training, or optimizing of network parameters, begins with an untrained CNN initialized randomly.

2.2. Proposed Method

Figure 1 depicts an overview of our proposed RWS-DIP method, in which the procedure of the target MR image reconstruction can be achieved in three steps: network training, MR image reconstruction, and data correction. In the first step, we do not need high-quality MR datasets and pre-training. The network parameters of the untrained CNN are optimized by solving the proposed constrained object function iteratively, which not only restricts the data consistency and explores wavelet sparsity but also introduces structural prior by using a similar reference image as the input of CNN. Next, the trained network outputs the reconstructed MR image. In the third step, the data correction process uses the prior measurements in -space to further improve the reconstruction accuracy. A further explanation will be provided in the following sections.

2.2.1. Network Training with a Reference and Wavelet Sparsity-Constrained DIP

Leveraging the concept of the traditional DIP framework, our proposed RWS-DIP method uses a high-resolution reference MR image and the wavelet sparsity to provide prior information for the target MR image reconstruction. Therefore, the objective function for network parameter optimization is as follows: where denote a high-resolution reference MR image acquired in advance with similar anatomical structure to the target image , is the wavelet transform operator, and is the norm. The regularization parameter .

Our proposed objective function in Equation (3) consists of the data fidelity term and the regularization term. It is aimed at finding the optimal network parameters that ensure the sparsity of the target MR image in wavelet domain on the premise of maintaining data consistency.

The data fidelity term restricts the data consistency between the network output and -space measurements. We use the known reference MR image as the network input, instead of random noise in traditional DIP. This strategy is capable of exploring and introducing the structural prior of the target MR image into the network for learning because of the high structural similarity between the reference and target images. The regularization constrains the sparsity of the target MR image in a wavelet domain, which merges more prior information for efficient training of network parameters.

Let , Equation (3) becomes

The constrained optimization problem in Equation (4) can be transformed into a penalty using the augmented Lagrangian:

In the expression above, stands for the Lagrange multiplier vector and is a penalty parameter.

To solve the problem in Equation (5), we use the alternating direction method of multipliers (ADMM) algorithm [42] to update the three unknowns , , and iteratively: (1)For the subproblem in Equation (6), this optimization is close in spirit to that performed in traditional DIP. However, we further modify the optimization by a proximity regularization that forces to be close to , which helps to provide additional stabilization and robustness(2)For the subproblem in Equation (7), the solution can be written as where is the soft thresholding operator defined as [42]

2.2.2. MR Image Reconstruction

After the iterative update procedure of network parameters, we obtain the trained CNN parameterized by (let MaxIt denote the maximum iteration number of ADMM; then, is the parameter of the final trained network). The output of the trained CNN is the reconstructed MR image, which can be presented as

2.2.3. Data Correction

Performing data correction operator to CNN output in the last step below, we obtain corrected -space data as follows: where denotes Fourier transform and is the priori acquired measurements of the target MR image, which are sampled at the spatial locations corresponding to the undersampled mask in -space. Let denote the complementary set of . This data correction strategy, defined in Equation (12), reserves all the priori acquired measurements to enforce the -space data consistency, so that the reconstruction error will focus only on the missing -space data. The final reconstructed target MR image can then be achieved by performing an inverse Fourier transform on

The algorithm flowchart of our proposed RWS-DIP method is presented in Algorithm 1.

Input:
  MaxIt - the iteration number for ADMM;
  SubIt - the iteration number for the update of network parameters;
   - the ADMM penalty parameter;
   - the regularization parameter;
   - the reference MR image;
   - the wavelet transform operator;
   - -space measurement;
Output: The reconstructed target MR image, ;
1: Initialization:, and set randomly;
2: =1: do
3: Update: Solve Eq.(6) using Adam and back-propagation for iterations;
4: Update: Apply Eq.(9) to obtain the solution of subproblem in Eq.(7);
5: Update:
6: end for
7: Reconstruction: CNN output ;
8: Data correction: ;
9: ;
2.3. Network Architecture

The CNN architecture employed in the proposed RWS-DIP method is summarized in Figure 1(b), which is the same as that used in [33]. It is an encoder-decoder (“hourglass”) architecture with skip connection. The encoding path (left side) and decoding path (right side) are linked by the skip connections, marked by yellow arrows, to integrate features from different resolutions. The network consists of repetitive applications of the convolutional (Conv) layer, batch normalization (BN) layer, and leaky rectified linear unit (LeakyReLU) layer, downsampling with stride and upsampling with bilinear interpolation. The maximal depth of the network is . , , and denote the number of filters at the th depth for downsampling, upsampling, and skip connections, respectively. , , and correspond to the respective kernel sizes.

3. Experimental Results

3.1. Experimental Setup

Experiments were conducted to evaluate the performance of our proposed RWS-DIP method. The comparisons with the proposed RWS-DIP method included zero-filling and traditional DIP [33]. To ensure a fair comparison, the zero-filling reconstructions and corresponding -space measurements were used as inputs for all the methods, and the same network architectures was employed for our RWS-DIP method and traditional DIP.

We quantified the reconstruction quality using the metrics of relative error (RelErr), peak signal-to-noise ratio (PSNR), and structural similarity index (SSIM) [43]:

In the descriptions in Equations (14)–(16), the reconstructed MR image and the ground truth are the same size of , and denotes the largest value in . Moreover, for the SSIM shown in Equation (16), , , , and represent the means and standard deviations of and , respectively, and denotes the crosscovariance between and , and constants and .

3.1.1. Data Acquisition

To demonstrate the performance of our RWS-DIP method, simulations were conducted on three groups of MR images. To simulate the data acquisition, we undersampled the 2D discrete Fourier transform of the MR images from MR scans, which were acquired from a 3T Siemens MRI scanner. The imaging parameters of the first group of scanned data (Brain A) were GR sequence, ,  ms,  ms, , and  mm. The reference and target images in Brain A were of size , as shown in Figures 2(a) and 2(b). The imaging parameters of the second and third groups of scanned data (Brain B and Brain C) were as follows: SE sequence, flip ,  ms,  ms, , and  mm. The MR images in Brain B and Brain C were of size and are shown in Figures 2(c)–2(f), respectively.

3.1.2. Training Setting

We used the same CNN architecture as the traditional DIP in [33], which is shown in detail in Figure 1(b). The parameters used in the experiments, including network hyperparameters, iteration number ( and ), wavelet (wavelet function and decomposition level), ADMM penalty parameter , and regularization parameter , are shown in Table 1.

The models were implemented on the Ubuntu 16.04 LTS (64 bit) operating system, running on an Intel Core i9-7920X 2.9 GHz CPU and Nvidia GeForce GTX 1080Ti GPU with 11 GB RAM in the PyTorch open framework with CUDA and CUDNN support.

3.1.3. Undersampled Schemes

To compare the influence of different undersampling masks to the performance of the proposed RWS-DIP method, our experiments employed three types of undersampling masks: Cartesian, variable density, and radial. Figure 3 depicts these three undersampling masks.

3.2. Results
3.2.1. Reconstruction Performance Comparison

(1) Reconstruction under Different Sampling Rates. We demonstrated the effectiveness of our RWS-DIP method at different sampling rates under Cartesian mask. Table 2 shows the quantitative performance of the proposed RWS-DIP method, traditional DIP and zero-filling reconstructions in RelErr, and PSNR and SSIM indexes at 10%, 20%, 30%, and 40% sampling rates. Taking into account the randomness involved in the training procedure (random initialization of network parameters in the proposed method; both random initializations of the network input and network parameters for traditional DIP), all the quantitative results were achieved by averaging the indices after being run 10 times. It can be seen that the proposed method has the lowest RelErr and the highest PSNR and SSIM values for all three groups of MR data, which means that our proposed RWS-DIP method can obtain more accurate reconstruction.

Figures 46 show the reconstructed MR images using the proposed RWS-DIP method and the compared methods under Cartesian undersampled mask with 20% and 30% sampling rates. It is obvious that our RWS-DIP method has the best performance in preserving more image textures and features, especially from the zoom-in images. The corresponding error images further show that the reconstruction of our RWS-DIP method has the smallest differences and is closest to the target MR image.

(2) Reconstruction with Different Undersampled Masks. The reconstruction results were compared under radial and variable density undersampled masks. The quantitative results tabulated in Table 3 clearly indicate that the proposed RWS-DIP method obtains more accurate reconstruction than with the radial and variable density undersampled masks. Comparisons of the reconstructed MR images are shown in Figures 7 and 8. The corresponding error images and zoom-in images demonstrate that our RWS-DIP method outperforms the compared methods with less structural loss and can preserve more details than the radial and variable density undersampled masks.

3.2.2. Convergence Analysis

Convergence is an important quality in applications of MRI methods based on deep learning. Therefore, we detected the convergence of the proposed RWS-DIP method use error curves drawn by conducting experiments on Brain A and Brain B under Cartesian undersampled mask. Figure 9 depicts the relative errors of reconstruction at every ADMM iteration. It can be observed that, as the number of iterations increases, the relative errors gradually converge to a low value at different sampling rates. Although there are slight fluctuations in the iteration procedure, the overall trend maintains convergence.

3.2.3. Parameter Evaluation

We evaluated the sensitivity of the proposed RWS-DIP method to parameter settings. The main parameters evaluated were the ADMM penalty parameter and the regularization parameter . We performed experiments on the Brain C dataset under Cartesian undersampled mask and varied one parameter at a time while keeping the rest as fixed values, as shown in Table 1.

Figures 10 and 11 show the plots of PSNR values as a function of the ADMM penalty parameter and the regularization parameter . As can be seen from the curves, the optimal numerical settings for and ( and ) in the proposed RWS-DIP method under different sampling rates are identical, which means that the RWS-DIP method has robustness in the setting of parameters. In fact, although the reconstructions have lower PSNR values than other numerical settings for parameters and , the difference is not significant, and the reconstruction performance is acceptable.

4. Conclusions

In this paper, we propose a novel reference-driven undersampled MR image reconstruction method using wavelet sparsity-constrained deep image prior. Our RWS-DIP method, which is based on the DIP framework, requires neither a pre-training procedure nor patient-based datasets, which is of great significance for clinical applications. The RWS-DIP method uses both structure and sparsity priors to improve the efficiency of the learning. The structural prior is introduced by employing a reference image as the network input, and the sparsity prior is explored by regularizing the norm of wavelet coefficients. Experimental results on MR scans show that the RWS-DIP method can achieve improved reconstruction performance and outperforms traditional DIP in preserving texture details and removing artifacts.

Two extensions can be made in order to improve the proposed scheme: (1) mining and incorporating more effective prior information may lead to a further boost in performance, particularly in regard to strengthening the use of structural prior information, and (2) further research is needed for the regularization effect introduced into DIP, which will guide the design of complementary regularizations, so as to achieve a stronger effect and better performance.

Data Availability

The data used to support the findings of this study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Acknowledgments

This work was supported in part by the National Natural Science Foundation of China under Grant 61527802, in part by the Key Science and Technology Project of Guangxi under Grant AB19110044, in part by the Guangxi Natural Science Foundation Innovation Research Team Project under Grant 2016GXNSFGA380002, in part by the Natural Science Foundation of Guangxi under Grant 2019GXNSFBA245076, in part by the Projects of Education Department of Guangxi under Grant 2020KY14016, in part by the Opening Foundation of Yulin Research Institute of Big Data under Grant 2020YJKY02, and in part by the Project of Yulin Normal University under Grant G2019ZK03.