Keywords

1 Introduction

Magnetic resonance imaging (MRI) is an imaging technique that can provide various contrast mechanisms for visualizing anatomical structures and physiological functions in human body. However, MRI is relatively slow because it is not possible to simultaneously sample multiple data points in the raw data domain (i.e. 2D Fourier domain of an image, which is referred to as k-space).

To accelerate MRI acquisition, k-space can be subsampled at a frequency that is lower than the Nyquist rate (i.e., it can be undersampled) instead of acquiring the fully-sampled k-space (i.e., sampling the data at the Nyquist rate). However, a simple reconstruction (i.e., 2D inverse Fourier transform) from the undersampled k-space brings out aliasing artifacts in the image domain and obscures many anatomical and physiological information. To reduce the aliasing artifacts and recover the missing information in images, various reconstruction algorithms such as compressed sensing (CS) [1] and parallel imaging (PI) [2] have been developed. The CS algorithms have been developed to the combination with low-rank constraint terms [3] and to image-adaptive algorithms that enforce sparsity on image patches [4].

In recent, deep-learning based reconstruction algorithms have been introduced and regarded as alternatives of CS. The first deep-learning algorithm applied to MRI reconstruction comprises a 3-layer convolutional neural network (CNN) that learns the relationship between undersampled images and fully-sampled images, and the conventional CS is followed at the end (hereafter denoted as Wang’s algorithm) [5]. After the study, various deep-learning algorithms such as CNN-based sparse residual (artifacts) learning algorithm [6], a cascaded CNN with interleaved data fidelity [7], multi-layer perceptron (MLP)-based parallel imaging (PI) algorithms have been introduced [8]. Especially, the automated transform by manifold approximation (AUTOMAP) algorithm provided a new perspective for reconstruction algorithm by directly translating k-space to the target image with neural networks [9].

In this study for accelerating MRI, we propose an algorithm that can efficiently translate 1D inverse Fourier transform (IFT) of undersampled k-space in data-acquisition direction (i.e., frequency-encoding direction) to the target fully-sampled images using two different deep-learning architectures: (1) a multi-layer perceptron (MLP) that estimates a target image from 1D inverse Fourier transform of k-space; and (2) a convolutional neural network (CNN) that reduces remaining artifacts in the output of the MLP. The proposed algorithm can utilize the maximum possible extent of the raw k-space while reducing the number of optimizing parameters dramatically compared with existing deep-learning algorithms, which enables efficient learning and results in better reconstructions than comparison algorithms.

2 Methods

2.1 Problem Formulation

Let \( {\mathbf{y}} \in {\complement }^{{{\text{n}}_{\text{kx}} \times {\text{n}}_{\text{ky}} }} \), where \( {\text{n}}_{\text{kx}} \) and \( {\text{n}}_{\text{ky}} \) represent the number of horizontal and vertical pixels of k-space respectively, denote a 2D complex-valued MR k-space. Our purpose is to reconstruct a fully-sampled image \( {\mathbf{x}} \in {\complement }^{{{\text{n}}_{\text{x}} \times {\text{n}}_{\text{y}} }} \), where \( {\text{n}}_{\text{x}} \) and \( {\text{n}}_{\text{y}} \) represent the number of horizontal and vertical pixels of x, from the undersampled k-space, \( {\mathbf{y}}_{{\mathbf{u}}} \), obtained by multiplication of a binary undersampling mask U and y as follows:

$$ {\mathbf{y}}_{\text{u}} = {\mathbf{U}} \circ {\mathbf{y}} = {\mathbf{U}} \circ {\mathcal{F}}_{2D} ({\mathbf{x}}) = {\mathbf{y}}_{{{\text{u}},{\text{r}}}} + {\text{i}}{\mathbf{y}}_{{{\text{u}},{\text{i}}}} $$
(1)
$$ {\mathbf{x}}_{\text{u}} = {\mathcal{F}}_{2D}^{ - 1} \left( {{\mathbf{y}}_{\text{u}} } \right) = {\mathbf{x}}_{{{\text{u}},{\text{r}}}} + {\text{i}}{\mathbf{x}}_{{{\text{u}},{\text{i}}}} $$
(2)

where \( {\mathbf{y}}_{{\mathbf{u}}} \in {\complement }^{{{\text{n}}_{\text{kx}} \times {\text{n}}_{\text{ky}} }} \) denotes the undersampled k-space; \( {\mathbf{U}} \in {\mathbb{R}}^{{{\text{n}}_{\text{kx}} \times {\text{n}}_{\text{ky}} }} \) denotes the binary undersampling mask; \( \circ \) denotes element-wise multiplication; \( {\mathcal{F}}_{2D} \) and \( {\mathcal{F}}_{2D}^{ - 1} \) denote the 2D Fourier transform (FT) and IFT, respectively; \( {\mathbf{y}}_{{{\mathbf{u}},{\mathbf{r}}}} \in {\mathbb{R}}^{{{\text{n}}_{\text{kx}} \times {\text{n}}_{\text{ky}} }} \) and \( {\mathbf{y}}_{{{\mathbf{u}},{\mathbf{i}}}} \in {\mathbb{R}}^{{{\text{n}}_{\text{kx}} \times {\text{n}}_{\text{ky}} }} \) denote the real and imaginary channels of \( {\mathbf{y}}_{{\mathbf{u}}} \), respectively; \( {\mathbf{x}}_{{\mathbf{u}}} \) denotes the undersampled image; and \( {\mathbf{x}}_{{{\mathbf{u}},{\mathbf{r}}}} \) and \( {\mathbf{x}}_{{{\mathbf{u}},{\mathbf{i}}}} \) denote the real and imaginary channels of \( {\mathbf{x}}_{{\mathbf{u}}} \), respectively.

To estimate x from a small number of samples in the k-space, \( {\mathbf{y}}_{{\mathbf{u}}} \), we introduce an algorithm comprising two deep-learning architectures, one is MLP that translates 1D IFT of undersampled k-space to an image, and the other is CNN that removes remaining artifacts in image domain. The objective function to optimize the parameters of the MLP, \( {\varvec{\Theta}}_{\text{MLP}} \), is

$$ {\text{argmin}}_{{{\varvec{\Theta}}_{\text{MLP}} }} \left\| {{\mathbf{x}} - H_{\text{MLP}} \left( {{\mathbf{z}}_{\text{u}} ;{\varvec{\Theta}}_{\text{MLP}} } \right)} \right\|_{2}^{2} $$
(3)

where \( {\mathbf{z}}_{\text{u}} = {\mathcal{F}}_{1D}^{ - 1} ({\mathbf{y}}_{\text{u}} ) \in {\complement }^{{{\text{n}}_{\text{x}} \times {\text{n}}_{\text{ky}} }} \) denotes the 1D IFT of \( {\mathbf{y}}_{\text{u}} \) along the frequency-encoding (horizontal) direction; and \( H_{\text{MLP}} \) denotes the hypothesis function that estimates fully-sampled image x from 1D IFT of undersampled k-space, \( {\mathbf{z}}_{\text{u}} \), by the MLP. The other objective function to optimize the parameters of the CNN for artifacts removal, \( {\varvec{\Theta}}_{\text{CNN}} \), is

$$ {\text{argmin}}_{{{\varvec{\Theta}}_{\text{CNN}} }} \left\| {{\mathbf{x}} - H_{\text{CNN}} \left( {{\hat{\mathbf{x}}}_{\text{MLP}} ;{\varvec{\Theta}}_{\text{CNN}} } \right)} \right\|_{2}^{2} +\uplambda\left\| {{\mathbf{y}}_{\text{u}} - {\mathbf{U}} \circ {\mathcal{F}}_{2D} \left( {H_{\text{CNN}} \left( {{\hat{\mathbf{x}}}_{\text{MLP}} ;{\varvec{\Theta}}_{\text{CNN}} } \right)} \right)} \right\|_{2}^{2} $$
(4)

where \( H_{\text{CNN}} \) denotes the hypothesis function for reducing the remaining artifacts in the image reconstructed by the MLP, \( {\hat{\mathbf{x}}}_{\text{MLP}} = H_{\text{MLP}} \left( {{\mathbf{z}}_{\text{u}} ;{\varvec{\Theta}}_{\text{MLP}} } \right) \); and \( \uplambda \) is the regularization parameter for data fidelity. The combined equation of Eqs. (3) and (4) to optimize \( {\varvec{\Theta}}_{\text{MLP}} \) and \( {\varvec{\Theta}}_{\text{CNN}} \), which is the final objective function of this study, can be represented as

$$ {\text{argmin}}_{{{\varvec{\Theta}}_{\text{MLP}} ,{\varvec{\Theta}}_{\text{CNN}} }} \begin{array}{*{20}c} {\left\| {{\mathbf{x}} - H_{\text{CNN}} \left( {H_{\text{MLP}} \left( {{\mathcal{F}}_{1D}^{ - 1} ({\mathbf{y}}_{\text{u}} );{\varvec{\Theta}}_{\text{MLP}} } \right);{\varvec{\Theta}}_{\text{CNN}} } \right)} \right\|_{2}^{2} } \\ { + \lambda \left\| {{\mathbf{y}} - {\mathbf{U}} \circ {\mathcal{F}}_{2D} \left( {H_{\text{CNN}} \left( {H_{\text{MLP}} \left( {{\mathcal{F}}_{1D}^{ - 1} ({\mathbf{k}}_{\text{u}} );{\varvec{\Theta}}_{\text{MLP}} } \right);{\varvec{\Theta}}_{\text{CNN}} } \right)} \right)} \right\|_{2}^{2} } \\ \end{array} $$
(5)

In Fig. 1, the proposed deep-learning frameworks to solve Eq. (5) are presented.

Fig. 1.
figure 1

The proposed reconstruction algorithm based on the MLP for translating 1D IFT of k-space to an image and the CNN for artifacts removal in the output of the MLP

2.2 The Proposed MLP: Translation from 1D IFT of K-space to an Image

The proposed MLP learns the relationship between 1D IFT of the undersampled Cartesian k-space which is transformed along the frequency-encoding (horizontal) direction, \( {\mathbf{z}}_{\text{u}} \), and the target fully-sampled image x. The proposed MLP is trained line by line rather than by a whole image, because each phase-encoding (vertical) line of the 1D IFT of k-space is not correlated with each other. Then, the estimation of a vertical line of the fully-sampled image x from the corresponding line of the 1D IFT of undersampled k-space, \( {\mathbf{z}}_{\text{u}} \), using the MLP can be represented as

$$ \widehat{{\mathbf{X}}}_{\text{MLP}} (\text{x}) = {\mathbf{w}}_{\text{NFL}} \left( { \ldots \sigma \left( {{\mathbf{w}}_{2} \sigma \left( {{\mathbf{w}}_{1} {\mathbf{Z}}_{\text{u}} ({\text{x}}) + {\mathbf{b}}_{1} } \right) + {\mathbf{b}}_{2} } \right) \ldots } \right) + {\mathbf{b}}_{\text{NFL}} $$
(6)

where

$$ \widehat{{\mathbf{X}}}_{\text{MLP}} ({\text{x}}) = \left[ {{\hat{\mathbf{x}}}_{\text{MLP}} ({\text{x}},1),{\hat{\mathbf{x}}}_{\text{MLP}} ({\text{x}},2), \cdots ,\varvec{ }{\hat{\mathbf{x}}}_{\text{MLP}} ({\text{x}},{\text{y}})} \right]^{{\mathbf{T}}} \in {\complement }^{{1 \times {\text{n}}_{\text{y}} }} $$
(7)
$$ {\mathbf{Z}}_{\text{u}} ({\text{x}}) = \left[ {{\mathbf{z}}_{\text{u}} ({\text{x}},1),{\mathbf{z}}_{\text{u}} ({\text{x}},2), \cdots ,{\mathbf{z}}_{\text{u}} \left( {{\text{x}},{\text{n}}_{{{\text{k}}_{\text{y}} }} } \right)} \right]^{{\mathbf{T}}} \in {\complement }^{{1 \times {\text{n}}_{\text{ky}} }} $$
(8)

where x and y denote horizontal and vertical indices in the image domain, respectively; \( {\mathbf{w}}_{\text{n}} \) and \( {\mathbf{b}}_{\text{n}} \) are weight and bias matrices of the MLP, respectively, where n = 1, 2, 3, …, NFL; \( \sigma \) is the activation function; and NFL denotes the number of fully-connected layers of the MLP. As depicted in Eq. (9), The loss function for training the MLP is defined as the mean-squared error between \( {\mathbf{Z}}({\text{x}}) \) and the estimation of \( {\mathbf{Z}}({\text{x}}) \) (i.e., \( H_{\text{MLP}} \left( {{\mathbf{Z}}_{\text{u}} ({\text{x}});{\varvec{\uptheta}}_{\text{MLP}} } \right) = \widehat{{\mathbf{X}}}_{\text{MLP}} ({\text{x}}) \)).

$$ L_{\text{MLP}} \left( {{\varvec{\uptheta}}_{\text{MLP}} } \right) = \frac{1}{{2{\text{M}}}}\sum\nolimits_{{{\text{m}} = 1}}^{\text{M}} {\left\| {{\mathbf{x}}\left( {{\text{x}}_{\text{m}} } \right) - H_{\text{MLP}} \left( {{\mathbf{Z}}_{\text{u}} \left( {{\text{x}}_{\text{m}} } \right);{\varvec{\uptheta}}_{\text{MLP}} } \right)} \right\|_{2}^{2} } $$
(9)

where

$$ {\varvec{\uptheta}}_{\text{MLP}} = \left\{ {\left( {{\mathbf{w}}_{1} ,{\mathbf{b}}_{1} } \right),\left( {{\mathbf{w}}_{2} ,{\mathbf{b}}_{2} } \right),\varvec{ } \ldots ,\left( {{\mathbf{w}}_{\text{NFL}} ,{\mathbf{b}}_{\text{NFL}} } \right)} \right\} $$
(10)

where \( {\mathbf{x}}\left( {{\text{x}}_{\text{m}} } \right) \) and \( {\mathbf{Z}}_{\text{u}} \left( {{\text{x}}_{\text{m}} } \right) \) denote the m-th vertical line of the fully-sampled image and the m-th vertical line of 1D IFT of the undersampled k-space, respectively; and M denotes the batch size. This line-by-line training can dramatically decrease the number of parameters to be learned because the number of input/output pixels (i.e., \( {\mathbf{Z}}_{\text{u}} ({\text{x}}) \) and \( \widehat{{\mathbf{X}}}_{\text{MLP}} ({\text{x}}) \)) decrease from \( {\text{n}}_{\rm{x}} {\text{n}}_{\rm{ky}} \) to \( {\text{n}}_{\rm{ky}} \). Finally, the hypothesis function and the parameters (i.e., weight and bias matrices) of the MLP that estimates the 2D fully-sampled image x from \( {\mathbf{z}}_{\text{u}} \) can be represented as

$$ H_{\text{MLP}} \left( {{\mathbf{z}}_{\text{u}} ;{\varvec{\Theta}}_{\text{MLP}} } \right) = {\hat{\mathbf{x}}}_{\text{MLP}} = {\mathbf{W}}_{\text{NFL}} \left( { \ldots \sigma ({\mathbf{W}}_{2} \sigma \left( {{\mathbf{W}}_{1} {\mathbf{z}}_{\text{u}} + {\mathbf{B}}_{1} } \right) + {\mathbf{B}}_{2} } \right) \ldots ) + {\mathbf{B}}_{\text{NFL}} $$
(11)

where

$$ {\varvec{\Theta}}_{\text{MLP}} = \left\{ {\left( {{\mathbf{W}}_{1} ,{\mathbf{B}}_{1} } \right),\left( {{\mathbf{W}}_{2} ,{\mathbf{B}}_{2} } \right),\varvec{ } \ldots ,\left( {{\mathbf{W}}_{\text{NFL}} ,{\mathbf{B}}_{\text{NFL}} } \right)} \right\} $$
(12)

where \( {\mathbf{W}}_{\text{n}} = 1{\mathbf{w}}_{\text{n}} \), \( {\mathbf{B}}_{\text{n}} = 1{\mathbf{w}}_{\text{n}} \), \( 1 = \left[ {1, 1, \cdots ,1} \right]^{{\mathbf{T}}} \in {\mathbb{R}}^{{1 \times {\text{n}}_{\text{ky}} }} \), and n = 1, 2, 3, …, NFL.

More details for the MLP, including optimizer parameters, network specifications, including network depths, the number of outputs, activation function, deep-learning libraries, and training/testing times, are provided in the supporting information.

2.3 The Proposed CNN with Data Fidelity: Artifacts Removal in the MLP Output

The proposed CNN aims to remove the artifacts remaining in the output images of the previous MLP. The hypothesis function of \( H_{\text{CNN}} \) that estimates the fully-sampled image x from the output of the MLP, \( {\hat{\mathbf{x}}}_{\text{MLP}} \), using CNN can be represented as

$$ H_{\text{CNN}} \left( {{\hat{\mathbf{x}}}_{\text{MLP}} ;{\varvec{\Theta}}_{\text{CNN}} } \right) = {\varvec{\upomega}}_{\text{NCL}} *\left( { \ldots \sigma ({\varvec{\upomega}}_{2} *\sigma \left( {{\varvec{\upomega}}_{1} *{\hat{\mathbf{x}}}_{\text{MLP}} + {\varvec{\upbeta}}_{1} } \right) + {\varvec{\upbeta}}_{2} } \right) \ldots ) + {\varvec{\upbeta}}_{\text{NCL}} $$
(13)

where

$$ {\varvec{\Theta}}_{\text{CNN}} = \left\{ {\left( {{\varvec{\upomega}}_{1} ,{\varvec{\upbeta}}_{1} } \right),\left( {{\varvec{\upomega}}_{2} ,{\varvec{\upbeta}}_{2} } \right),\varvec{ } \ldots ,\left( {{\varvec{\upomega}}_{\text{NCL}} ,{\varvec{\upbeta}}_{\text{NCL}} } \right)} \right\} $$
(14)

where \( {\varvec{\upomega}}_{\text{n}} \) and \( {\varvec{\upbeta}}_{\text{n}} \) are convolution and bias matrices in each convolution layer, respectively, where \( {\text{n}} = 1,2,3, \ldots ,{\text{NCL}} \); \( * \) denotes the convolution operator; and NCL denotes the number of convolution layers. The loss function used for training the CNN is

$$ L_{\text{CNN}} \left( {{\varvec{\Theta}}_{\text{CNN}} } \right) = \frac{1}{{2{\text{M}}}}\sum\nolimits_{{{\text{m}} = 1}}^{\text{M}} {\left\| {{\mathbf{x}}_{\text{m}} - H_{\text{CNN}} \left( {{\hat{\mathbf{x}}}_{{{\text{MLP}},{\text{m}}}} ;{\varvec{\Theta}}_{\text{CNN}} } \right)} \right\|_{2}^{2} } $$
(15)

where \( {\mathbf{x}}_{\text{m}} \) and \( {\hat{\mathbf{x}}}_{{{\text{MLP}},{\text{m}}}} \) denote the m-th fully-sampled image and the m-th estimated image of the MLP in the training dataset, respectively; and M denotes the batch size.

The final data fidelity is performed in k-space as a closed form solution [4] as

$$ {\hat{\mathbf{y}}}_{\text{f}} \left( {{\text{k}}_{\text{x}} , {\text{k}}_{\text{y}} } \right) = \left\{ {\begin{array}{*{20}c} {\frac{{{\hat{\mathbf{y}}}_{{{\mathbf{MLP}} + {\mathbf{CNN}}}} \left( {{\text{k}}_{\text{x}} , {\text{k}}_{\text{y}} } \right) +\uplambda{\mathbf{y}}_{{\mathbf{u}}} \left( {{\text{k}}_{\text{x}} , {\text{k}}_{\text{y}} } \right)}}{{1 +\uplambda}}} & {{\text{if}}\;{\mathbf{U}}\left( {{\text{k}}_{\text{x}} ,{\text{k}}_{\text{y}} } \right) = 1} \\ {{\hat{\mathbf{y}}}_{{{\mathbf{MLP}} + {\mathbf{CNN}}}} \left( {{\text{k}}_{\text{x}} , {\text{k}}_{\text{y}} } \right)} & { {\text{if}}\;{\mathbf{U}}\left( {{\text{k}}_{\text{x}} ,{\text{k}}_{\text{y}} } \right) = 0} \\ \end{array} } \right. $$
(16)

where

$$ {\hat{\mathbf{y}}}_{{{\mathbf{MLP}} + {\mathbf{CNN}}}} = {\mathcal{F}}_{2D} \left( {\hat{x}_{{{\mathbf{MLP}} + {\mathbf{CNN}}}} } \right) = {\mathcal{F}}_{2D} \left( {H_{\text{CNN}} \left( {{\hat{\mathbf{x}}}_{\text{MLP}} ;{\varvec{\Theta}}_{\text{CNN}} } \right)} \right) $$
(17)

where \( {\text{k}}_{\text{x}} \) and \( {\text{k}}_{\text{y}} \) are horizontal and vertical indices of k-space, respectively. Then, the final solution of the proposed algorithm is obtained by IFT of \( {\hat{\mathbf{y}}}_{\text{f}} \) as \( {\hat{\mathbf{x}}}_{\text{f}} = {\mathcal{F}}_{2D}^{ - 1} \left( {{\hat{\mathbf{y}}}_{\text{f}} } \right) \).

More details for the CNN, including optimizer parameters, network specifications, including network depths, filter sizes, activation function, deep-learning libraries, and training/testing times, are provided in the supporting information.

2.4 Experimental Frameworks

All experiments conducted in the present study were approved by the institutional review board. Written informed consent was obtained from all human subjects. Two MR datasets were used: T2-fluid attenuated inversion recovery (T2-FLAIR) brain real-valued dataset provided by the Alzheimer’s Disease Neuroimaging Initiative (ADNI) [10], and T2-weighted complex-valued dataset acquired at our local institute. Details of data acquisition, including scanner information, sequence parameters, and the number of slices used for training and testing, are provided in the supporting information.

Undersampled k-space data were retrospectively obtained by subsampling the fully-sampled k-space data. Before undersampling, all MR images were normalized to a maximum magnitude of 1. A Cartesian undersampling scheme in a phase encoding direction (1D random) was used for generating undersampled k-space dataset. The sampling ratio for the undersampling was 25%. The binary undersampling masks is presented in the supporting information (Supporting Figure S1).

The proposed algorithm was compared with the following six algorithms: baseline zero-filling, CS-MRI [1], DL-MRI [4], a CNN-based algorithm by Wang et al. [5] (denoted as Wang’s algorithm), PANO [11], and FDLCP [12]. The parameters of the comparison algorithms are provided in the supporting information. Peak signal-to-noise ratio (PSNR) and structure similarity (SSIM) [13] were used for quantitative metrics. The patch size used to calculate SSIM was 11.

3 Results

3.1 The Conventional Image-to-Image MLP vs. the Proposed MLP

In Fig. 2, we evaluated the two different MLPs: The conventional MLP introduced in [8] which translates each vertical line of the 2D IFT of the undersampled k-space (i.e., the undersampled image) to the corresponding line of the target image and the proposed MLP which translates each vertical line of 1D IFT of the undersampled k-space to the corresponding line of the target image. Variables such as the number of outputs and the network depths in MLPs were exactly same. In Fig. 2, the proposed MLP (h) showed better reconstructions than the conventional image-to-image MLP (g) in terms of restoring details and reducing aliasing artifacts.

Fig. 2.
figure 2

(a) the fully-sampled image in T2-FLAIR (ADNI) dataset, (b) the zero-filling image at the sampling ratio = 25%, the reconstructed images by (c) the image-to-image MLP and (d) the proposed MLP. (e–h) are the magnified images of boxed region of interests (ROIs) in (a–d).

3.2 Comparison Study with Conventional Algorithms

Qualitative and quantitative comparisons are provided for the six conventional algorithms and the proposed algorithm. Figure 3 depicts the resultant images from the complex-valued T2-weighted dataset (from our local institute). The results of the real-valued T2-FLAIR dataset are provided in the supporting information (Supporting Figure S3). Table 1 depicts quantitative results for the two data sets. The proposed algorithm outperformed the six comparison algorithms in both qualitative and quantitative tests.

Fig. 3.
figure 3

(a) the fully-sampled image from the T2-weighted (our local institute) dataset, (b) zero-filling image with the sampling ratio = 25%, and the images reconstructed with (c) CS-MRI, (d) DL-MRI, (e) Wang’s algorithm, (f) PANO, (g) FDLCP, and (h) the proposed algorithm. (i–p) are the magnified images of the boxed ROIs in (a–h).

Table 1. Quantitative results of comparison algorithms and the proposed algorithm. Average PSNR/SSIMs for the sampling ratio = 25% on the two datasets (real-valued T2-FLAIR and complex-valued T2-weighted)

4 Conclusion

The current study presents a domain-transform algorithm comprising two deep-learning architectures (MLP and CNN) for reconstructing MR images from undersampled Cartesian k-space. Experimental results on the two different datasets demonstrated that the proposed algorithm outperforms the comparison state-of-the-art reconstruction algorithms. More discussion is provided in the supporting information.