 Research
 Open Access
 Published:
Compressed sensing MRI via fast linearized preconditioned alternating direction method of multipliers
BioMedical Engineering OnLine volume 16, Article number: 53 (2017)
Abstract
Background
The challenge of reconstructing a sparse medical magnetic resonance image based on compressed sensing from undersampled kspace data has been investigated within recent years. As total variation (TV) performs well in preserving edge, one type of approach considers TVregularization as a sparse structure to solve a convex optimization problem. Nevertheless, this convex optimization problem is both nonlinear and nonsmooth, and thus difficult to handle, especially for a largescale problem. Therefore, it is essential to develop efficient algorithms to solve a very broad class of TVregularized problems.
Methods
In this paper, we propose an efficient algorithm referred to as the fast linearized preconditioned alternating direction method of multipliers (FLPADMM), to solve an augmented TVregularized model that adds a quadratic term to enforce image smoothness. Because of the separable structure of this model, FLPADMM decomposes the convex problem into two subproblems. Each subproblem can be alternatively minimized by augmented Lagrangian function. Furthermore, a linearized strategy and multistep weighted scheme can be easily combined for more effective image recovery.
Results
The method of the present study showed improved accuracy and efficiency, in comparison to other methods. Furthermore, the experiments conducted on in vivo data showed that our algorithm achieved a higher signaltonoise ratio (SNR), lower relative error (Rel.Err), and better structural similarity (SSIM) index in comparison to other stateoftheart algorithms.
Conclusions
Extensive experiments demonstrate that the proposed algorithm exhibits superior performance in accuracy and efficiency than conventional compressed sensing MRI algorithms.
Background
Magnetic resonance imaging (MRI), which is noninvasive, provides nonelectromagnetic radiation, higher softtissue contrast, and spatial resolution, has been applied in diagnostic medicine for many years. However, due to the limitations of the hardware scanning system and the traditional Nyquist sampling theory, MRI scanners take a considerable length of time to acquire kspace data. Patient motion (e.g. a beating heart and respiratory movement) during lengthy scans can cause motion and streaking artifacts on the reconstructed image. This degrades image quality, which could lead to misdiagnosis. Thus, accelerating the sampling speed and reducing or eliminating artifacts have always been the aims of many studies.
With the rapid development of the novel compressed sensing (CS) theory [1, 2], compressed sensing MRI (CSMRI) has attracted much attention, as it can reduce imaging time considerably. Compressed sensing theory claims that by using random projection, a small number of data points can be directly sampled at a sampling frequency that is far below the Nyquist frequency. In CSMRI, the imaging time can be significantly reduced by reconstructing an image of good quality from highly undersampled kspace data. Specifically, if the signal or image is sparse in a certain domain, we can obtain perfect reconstruction with sufficient measurements. MR images are generally sparse in some transform domain, such as the wavelet domain. Consequently, the CS technique can be easily combined with MRI. The original signals or images can be recovered by using the nonlinear reconstruction algorithm under the restricted isometry property (RIP) [3, 4]. Nevertheless, owing to the limitations of MR physics, MRI cannot achieve twodimentional random sampling.
In 2007, Lustig et al. [5] proposed the SparseMRI algorithm, which selects wavelet transform as a sparse basis, and uses variable density random sampling and the conjugate gradient descent method for image recovery. This was the first application of CS to MRI. However, due to a high time complexity, the SparseMRI is too slow to be put into practical use. Since then, a variety of nonlinear algorithms have been proposed for CSMRI reconstruction. The alternating direction method of multipliers (ADMM) algorithm has been studied extensively [6,7,8], and has been widely used in optimization problems that arise in machine learning, image processing, etc. Recently, the fast alternating direction method of multipliers (FADMM) [9] has incorporated a predictorcorrector acceleration scheme into the simple ADMM, when a strongly convex condition is satisfied. This algorithm cannot guarantee a global convergence when weakly convex problems are encountered. Another fast method, referred to as the accelerated alternating direction method of multipliers (ALPADMM) [10], was proposed to deal with a class of affine equality constrained composite optimization problems. Although ALPADMM is capable of handling saddle point problems, its convergence rate largely depends on the Lipschitz constant of the smooth component.
In this article, we propose an efficient algorithm to solve an augmented total variation (TV)regularized model that adds a quadratic term to the classical TVregularized model to enforce smoothness of the image. The proposed method applies a linearization strategy to two quadratic terms and divides the original convex problem into two subproblems, both of which are easy to solve. For all subproblems, the augmented Lagrangian function, which combines both the Lagrangian function and the quadratic penalty function, is applied to update each variable and gain more reconstruction accuracy than the Lagrangian function approach alone. Furthermore, we utilize a multistep weighted technique to improve the accuracy of reconstruction. Numerical experiments have been conducted to compare the proposed algorithm with previous algorithms on various MR images. The experimental results show that the proposed approach can achieve a higher signaltonoise ratio (SNR), lower relative error (Rel.Err), and better structural similarity (SSIM) index. The main contributions of the work are twofold as follows: (i) the proposed linearized preconditioned alternating direction method of multipliers (FLPADMM) that is inspired by the smooth technique [9], linearized strategy [10, 11], and the accelerated method [10] is designed to solve the augmented TVregularized model; and (ii) this algorithm only linearizes the closed convex function and does not require the application of multistep weighting to each variable.
The paper is organized as follows: In "Related work", the CSMRI reconstruction algorithms are reviewed. "Methods" briefly describes the basics of CS problem formulation and the proposed FLPADMM method to reconstruct MR images. The experimental results of the proposed approach and comparison with other algorithms are illustrated in "Results". Corresponding discussions are given in "Discussion" and conclusions are provided in "Conclusions".
Related work
In this section, we briefly review the conventional CSMRI reconstructed algorithms. Many nonlinear algorithms, e.g., the iterative shrinkage/thresholding method (IST) [12], twostep IST (TwIST) scheme [13], fast IST algorithm (FISTA) [14], split augmented Lagrangian shrinkage algorithm (SALSA) [6], wavelet tree sparsity MRI (WaTMRI) [15], total variation augmented Lagrangian alternating direction method (TVAL3) [16], total variation based compressed magnetic resonance imaging (TVCMRI) [17], reconstruction from partial Fourier data (RecPF) [18], and fast composite splitting algorithm (FCSA) [19], have been proposed to improve reconstruction speed and accuracy. IST is an operatorsplitting algorithm that can be applied to an optimization problem with a simpler regularization term. The global acceleration of IST may be very slow especially when the stepsize is quite small or the optimization problem is extremely illconditioned. TwIST, which is a variant of IST, utilizes two or more previous iterates to update the current values, and does not depend on the previous iterate alone. TwIST gains higher speed than IST on reconstruction problem; however, the global convergence rate of TwIST has not been thoroughly studied. FISTA is another accelerated variant of IST that also takes advantage of two previous iterates. Unlike TwIST, FISTA can achieve global convergence with a splitting scheme.
SALSA transforms a constrained problem into an unconstrained problem by adding a penalty term that can be dealt with, using an augmented Lagrangian approach. Many image linear inverse problems, including image deblurring, MRI image reconstruction, and image inpainting, can be handled with SALSA. According to structured sparsity theories, WaTMRI exploits the tree sparsity to improve CSMRI, which combines standard wavelet sparsity with total variation. To our knowledge, TVAL3 was proposed to solve a set of equalityconstrained nonsmooth problems, and integrated an alternative minimization technique with a nonmonotone line search to optimize an augmented Lagrange function. Both TVCMRI and RecPF utilize a splitting strategy to minimize objective function. The former uses operatorsplitting technology, and the latter adopts a variable splitting method to obtain optimal solution. In FCSA, the original optimization problem is divided into two easier subproblems that can be easily solved by FISTA. However, these algorithms are not necessarily easily implemented, or even sufficient to solve CS problems with TVregularization, especially when the measurement matrix is considerably illconditioned. Therefore, it is necessary to develop algorithms that are both accurate and efficient to solve largescale problems.
Methods
Problem formulation
Generally, the classical TVregularized model for CSMRI reconstruction problems can be written as:
where \(x \in R^{n}\) is the image to be reconstructed, \(b \in R^{m}\) denotes the undersampled kspace data from the MR scanner, and \(A\in R^{m \times n} (m < n)\) is the measurement matrix. The expression \(\delta > 0\) represents the noise level in the measurement data and \((\nabla x)_{i,j} = (x_{i+1,j}x_{i,j},x_{i,j+1}x_{i,j})\), where \(\nabla\) is the discrete gradient operator and \(\nabla x\) denotes the TV regularization of x. A variant of (1) is the following TVregularization problem:
where \(\tau > 0\) is a positive regularization parameter to balance between the two objectives. To enforce image smoothness, we add a quadratic term \(\frac{\gamma }{2}\Vert \nabla x\Vert ^2_{2}\) in the objective function to give the new TVregularized model:
where \(\gamma\) is a smoothing parameter. The augmented term \(\nabla x + \frac{\gamma }{2}\Vert \nabla x\Vert ^2_{2}\) can yield accurate solutions by using a proper value of \(\gamma\). In addition, the dual problem is continuously differentiable and facilitates effective use of gradient information. To split the variable x, an auxiliary variable z is introduced by \(\nabla x  z = 0\), and the unconstrained optimization problem (3) is transformed into:
The augmented Lagrangian function for problem (4) is given by:
where \(\uplambda \in R^{m}\) is the Lagrangian multiplier, \(\mu\) is a positive penalty parameter, and \(\langle\uplambda , \nabla x z\rangle\) denotes the inner product of the vectors \(\uplambda\) and \(\nabla xz\). The classical ADMM minimizes the convex optimization problem (5) with respect to x and z, using the nonlinear block–Gauss–Seidel technique. After minimizing x and z alternatively, \(\uplambda\) can be updated by:
The nonlinear block–Gauss–Seidel iteration of ADMM can be written as:
Suppose \(z_{k}\) and \(\uplambda _{k}\) are given, then \(x_{k+1}\) can be obtained by:
When \(x_{k+1}\) and \(\uplambda _{k}\) remain fixed, \(z_{k+1}\) can be minimized by:
The ADMM algorithm above, used to solve (4), is expressed in Algorithm 1.
The ADMM has been previously studied and analyzed [7, 8, 20]. Generally, if subproblems in (7) are not in closedform, many solutions could exist within those subproblems. Moreover, when the objective functions are poor or difficult to handle at high precision, the conventional ADMM algorithms might also perform poorly in image reconstruction.
Proposed algorithm
Based on the above analysis, firstly, the minimization of (8) is given by:
Hence, (10) is transformed into:
Because the measurement matrix A is neither the identity matrix, nor is it typically fully dense, it is impossible to derive the exact solution with respect to the x vectors [21]. In addition, the computational cost to handle (11) is extremely heavy. In order to reduce the computational burden and get closedform solutions, some variants could be considered. The linearization of the quadratic term \(\frac{1}{2} \Vert Axb\Vert ^2_{2}\) is used to update \(x_{k+1}\) as follows:
where \(Grad(x_{k}) = A^T(Ax_{k} b)\) is the gradient of \(\frac{1}{2} \Vert Axb\Vert ^2_{2}\) at the current point \(x_{k}\), and \(\eta\) is a positive proximal parameter. Then, the x subproblem in (10) can be iterated by:
Considering the quadratic term \(\frac{\mu }{2}\Vert \nabla x  z_{k}\Vert ^2_{2}\) can also be linearized, we also linearize \(\frac{\mu }{2}\Vert \nabla x  z_{k}\Vert ^2_{2}\) at \(x_{k}\). This variant is a fast linearized preconditioned ADMM(FLPADMM) algorithm that generates the iterates \(x_{k+1}\) by:
The negative divergence operator \(div\) can be used to solve (14) as follows:
Secondly, for a given \(x_{k + 1}\) and \(\uplambda _{k}\), \(z_{k+1}\) is computed by solving:
Hence, the solution to (16) obeys:
where \(soft( \cdot , T)\) is the soft thresholding function that is defined as:
Finally, the Lagrangian multiplier is updated by \(\uplambda _{k+1}=\uplambda _{k}  \mu (\nabla x_{k+1}z_{k+1})\).
The FLPADMM algorithm utilizes the gradient descent method and one soft thresholding operator to update variables at each iteration. In addition, this method is a variant of the classical ADMM algorithm framework. The proposed FLPADMM algorithm is presented in Algorithm 2.
The stopping criterion in all of the algorithms above is the relative change of x between two successive iterations, which is small enough, i.e.:
where tol is usually a range chosen from \(10^{5}\) to \(10^{3}\). In the FLPADMM algorithm, \(\alpha _{k}\) represents a weighted parameter and \(Grad(x^{m}_{k})\) is the gradient of \(\frac{1}{2} \Vert Axb\Vert ^2_{2}\) at the point \(x^{m}_{k}\). Furthermore, \(x^{m}_{k}\) is the first weighted value that is used to update \(x_{k+1}\).
It should be noted that the appropriate parameter \(\alpha _{k}\) can improve the rate of convergence that has been proven in a previous study [10]. Moreover, the superscript m stands for “middle,” and w stands for “weight.” Before the gradient descent method in line 5 is applied, x is updated by the weighted sums of all previous iterates. Furthermore, after the gradient method is applied, x is updated again by the same weighting technique, that is, x is weighted twice at each iteration. Specifically, when the weighted parameter \(\alpha _{k}\) is set to 1, the x subproblem is simply the current point \(x_{k+1}\). At this point, FLPADMM becomes another variant of the ADMM algorithm [10]. The accelerated strategy of FLPADMM incorporates a multistep acceleration scheme with middle point \(x^{m}\) and weighted point \(x^{w}\), which was first applied in a previous study [10] and derived from the accelerated gradient method [22, 23]. Moreover, the optimal rate of convergence of FLPADMM is \(O\left(\frac{1}{k^{2}} + \frac{1}{k}\right)\).
Results
Experimental setup
A series of numerical experiments were conducted to compare the performance of the proposed FLPADMM with two stateoftheart algorithms, namely FADMM [9] and ALPADMM [10], for MR image reconstruction from undersampled kspace data. The four typical MR datasets (Shepp–Logan phantom, human brain1 MR data, human brain2 MR data, and human spine MR data) were used to evaluate our algorithm. All test images had the same matrix size of 256 × 256, as shown in Fig. 1a–d. The first Shepp–Logan phantom was a piecewise smooth image with pixel intensities ranging from 0 to 1. The complex kspace data of the human brain1 was acquired from a 3T GE MR750 scanner using the FRFSE sequence (TR = 6000 ms, TE = 101 ms). The human brain2 data was also obtained from the 3T GE MR750 system (TR/TE = 2500/96.9 ms, field of view = 280 × 280 mm, slice thickness = 5 mm). The human spine MR data was fully sampled kspace data acquired on a 3T GE MR750 system with a FRFSE sequence (TR/TE = 2500/110 ms, field of view = 240 × 240 mm, slice thickness = 5 mm). To achieve fair comparisons, codes of all compared algorithms were downloaded from the authors’ websites. All experiments were executed, using Windows 10 and MATLAB 2015b (64bit), on a desktop computer with a 3.2GHz Intel core i54460 CPU and 8GB of RAM.
Each experiment was repeated 10 times, and the average image metric results of 10 experiments were recorded. For most of the MR images, the kspace signals with a large magnitude were generally localized in the central region. Since a nonCartesian sampling matrix is incoherent, the results on the Cartesian masks were far less favorable than those on the nonCartesian mask. Therefore, two nonCartesian masks were chosen as the sampling masks. One was a pseudoGaussian mask, displayed in Fig. 2a, that was implemented by following the sampling strategy of collecting more lowfrequency signals in the central part of the kspace, and less highfrequency signals in the peripheral part of the kspace. The other mask presented in Fig. 2b, was a pseudoradial mask that was applied by following the rule of RecPF [18]. The sampling ratio was defined as the number of sampled points divided by the total size of the original image.
In the present study, we compared our algorithm with two stateoftheart algorithms under similar conditions. To explore the influence of the regularization parameter \(\tau\), we used human brain1 data as an example, to analyze the changes of image quality when \(\tau\) was changed. In Fig. 3, the SNR attained the maximum value when \(\tau\) was 10^{−3}. Thus, we chose this optimum value to achieve favorable reconstruction. Similar searches were adopted for the other datasets. For all tests, we also found that when \(\gamma = 2\tau\), our algorithm maintained the most favorable reconstruction performance. Furthermore, the default maximum of all three methods was set to 300.
To quantitatively evaluate the result of the proposed algorithm, three objective metrics were adopted to measure the quality of the recovered images. The first was SNR, defined as:
where x is the original image, \(\overset{\wedge }{x}\) is the reconstructed image, and M and N represent the number of rows and columns, respectively, in the input image. The quality of the reconstructed image is directly proportional to the value of SNR. The second metric was the Rel.Err, defined as:
A smaller value meant that the reconstructed image had little error and more favorable reconstruction in comparison to the original image.
The last metric was the SSIM index that was used to measure the similarity between two images, in terms of structure, brightness, and contrast, among other aspects, and defined as:
where \(\mu _{p}\) and \(\theta _{p}\) are the mean and variance, respectively, of the original image; \(\mu _{q}\) and \(\theta _{q}\) are the mean and variance, respectively, of the reconstructed image, \(\theta _{pq}\) is the covariance of these two images; and \(c_{1}\) and \(c_{2}\) are fixed constants that prevent unstable phenomena when the denominator is close to zero. When the value of SSIM was increased, the image showed greater similarity to the original.
Experimental results
In this section, we first compare our proposed FLPADMM algorithm with FADMM [9] and ALPADMM [10] algorithms on the Shepp–Logan phantom, with Gaussian white noise of a standard deviation of 0.01. The proposed FLPADMM was applied to the Shepp–Logan phantom under pseudoGaussian mask with 20% kspace data undersampled. Figure 4a shows the original Shepp–Logan phantom, and Fig. 4b–d presents the reconstructed images recovered by the FADMM, ALPADMM, and FLPADMM algorithms, respectively. Compared with the original Shepp–Logan phantom image, FADMM yielded noticeable artifacts and failed to suppress background noise. The image recovered by ALPADMM contained fewer artifacts and was evidently more favorable than that recovered by FADMM. As the Shepp–Logan phantom is extremely piecewise smooth and sparse, ALPADMM also provides good reconstruction. As shown in Fig. 4c, d, visible artifacts are not easily observed when both ALPADMM and FLPADMM are used. However, for experiments on in vivo data as we will show, FLPADMM would perform much more accurately and stably than ALPADMM.
For enhanced visualization, Fig. 4e–g depicts the difference between the reconstructed image and the original image of the Shepp–Logan phantom under a pseudoGaussian mask at a sampling ratio of 20% using FADMM, ALPADMM, and the proposed FLPADMM reconstruction. It was evident that the reconstruction with FLPADMM had the smallest error. The proposed FLPADMM exhibited superior performance in suppressing noise without significant artifacts, and yielded the best reconstruction.
All experiments on in vivo data were corrupted with Gaussian white noise with zero mean and a standard deviation of 0.01. Experimental results of these in vivo human brain data are displayed in Figs. 5 and 6 at a sampling ratio of 25%. Figure 5a presents the original human brain1 image. The reconstructed images illustrated in Fig. 5b–d, were obtained by FADMM, ALPADMM, and our proposed FLPADMM, under a pseudoGaussian sampling scheme. Figure 5e–g was produced by FADMM, ALPADMM, and FLPADMM, respectively, under a pseudoradial sampling pattern. The SNR of the human brain1 image under the pseudoGaussian mask using FLPADMM was 25.0685 dB, whereas those recovered by FADMM and ALPADMM were 20.9921 and 22.3231 dB, respectively. We can clearly see that the FLPADMM reconstruction suppressed background noise. The recovery result in Fig. 6 is similar to that in Fig. 5.
Figure 7 gives the comparison results of human brain1 data among the stateoftheart MR image reconstruction algorithms, using different sampling masks, when the sampling ratios were 0.1, 0.2, 0.3, 0.4, and 0.5, respectively. As seen in Fig. 7a, b, the proposed FLPADMM achieved high image quality with high SNR and low Rel.Err. When the sampling ratio was 0.1, the three methods performed relatively poorly. That is because when sampling ratio is too low, the sampled data is insufficient to obtain a faithful image. It is notable that as the sampling ratio increased for all algorithms under consideration, the SNR was also increased, whereas Rel.Err was gradually reduced. That is, the reconstructions of higher quality could have been obtained by taking more measurements. In addition, when the sampling ratio was increased, the FLPADMM algorithm exhibited superior performance in recovering the sampling image. Specifically, a sampling ratio of 30% was sufficient to reconstruct the human brain1 image effectively.
Figure 8 shows the comparison results of human brain2 data, where (a) and (c) present the SNR of the human brain2 image with different ratios, and (b) and (d) describe the Rel.Err of the human brain2 image at different ratios. The results are similar to those of the human brain1 image. The SNR of FLPADMM was slightly larger than that of ALPADMM when a pseudoradial mask was applied. The Rel.Err of ALPADMM was very close to that of FLPADMM (see Figs. 7d, 8d), indicating that both ALPADMM and our proposed FLPADMM show similar recovery performance under the pseudoradial mask.
The reconstructed results in Fig. 9 were consistent with those of Fig. 5. Compared to FADMM and ALPADMM, our proposed FLPADMM reconstructed better images without visual artifacts. For example, when the sampling ratio was 25% under pseudoradial sampling, FADMM had significant artifacts and ALPADMM had slight artifacts. However, images reconstructed by FLPADMM were the closest to the original image of the human spine. These results further validate the superiority of FLPADMM in comparison to other algorithms and are consistent with the results of the two human brain experiments. It is clear that regardless of the sampling scheme, FLPADMM achieved the highest SNR and lowest Rel.Err. From Fig. 10, we can see that all reconstruction results showed steady improvement as the sampling ratio increased. Moreover, FLPADMM showed superior performance in comparison to other algorithms. As the tissue structure of the human spine MR data was extremely complex, the quality of the reconstructed image was not as good as that of the human brain tests.
For further comparison, the results of quantitative image metrics, SSIM and CPU time (s), are listed in Tables 1 and 2 to demonstrate structural similarity and running time of FADMM, ALPADMM, and FLPADMM algorithms for all MR images at various sampling ratios. As seen in Table 2, the running time of FLPADMM was faster than that of ALPADMM, but slower than that of FADMM by approximately 1 s. Since the convergence of FADMM was also close to \(O(\frac{1}{k^2})\) when strict conditions were satisfied, the running time of all three methods were very similar. According to the value of the SSIM, the proposed FLPADMM achieved higher quality images than the other techniques.
Discussion
We proposed a novel algorithm for CSMRI reconstruction, referred to as FLPADMM. Except for the TVregularization term in the classical MR model, we added a quadratic term to this classical model to make the image smoother. Using augmented Lagrangian function, FLPADMM effectively divides the original convex problem into two subproblems, both of which can be easily dealt with. To further enhance image reconstruction, a strategy that incorporated a multistep weighted scheme was adopted in FLPADMM. Several parameters need to be tuned in our proposed algorithm. In general, the requirement on stepsize \(\eta\) obeys \(\eta > \mu \Vert A^{T}A\Vert\). When the regularization parameter \(\tau\) is \(10^{3}\) (see Fig. 3) and \(\gamma = 2\tau\), our proposed algorithm yields the best result. The other parameters can be manually set for different test data under a fixed sampling scheme. When different sampling schemes (i.e., pseudoGaussian mask and pseudoradial mask) are applied, our proposed FLPADMM can also produce very impressive results. Experiments validate that the performance of this proposed method is superior to those of FADMM and ALPADMM. It particularly shows the best performance in suppressing background noise, even at a low sampling ratio.
Some algorithms that combine parallel MRI and CS have been proposed to accelerate MRI reconstruction [24,25,26]. Our method can also be applied to parallel MRI with minor revisions.
Conclusions
The consideration of TVregularization for CSMRI has been studied within recent years, largely because MR images can be recovered from its partial Fourier samples, and TV shows better performance in preserving image edges. In this paper, we first briefly reviewed nonlinear algorithms for CSMRI, and then introduced an augmented TVregularized model with an additional quadratic term to enforce image smoothness. An efficient, inexact, but unique algorithm has been proposed to handle this novel TVregularized model. The proposed algorithm, referred to as FLPADMM, belongs to the classical ADMM framework that decomposes the objective function into two subproblems by adding new variables and constraints. FLPADMM minimizes the TVregularized objective function by an augmented Lagrangian minimization function technique. Furthermore, this method effectively adopts a multistep weighted scheme to improve the accuracy of reconstruction. Moreover, FLPADMM could also solve both constrained and unconstrained convex optimization problems. Numerous experiments demonstrate the superiority of the proposed FLPADMM method in comparison to the previous FADMM and ALPADMM algorithms. Our future work would combine this algorithm with parallel MRI to further accelerate the imaging time.
Abbreviations
 MRI:

magnetic resonance imaging
 CS:

compressed sensing
 ADMM:

alternating direction method of multipliers
 TV:

total variation
 RIP:

restricted isometry property
 IST:

iterative shrinkage/thresholding algorithm
 TwIST:

twostep iterative shrinkage/thresholding algorithm
 FISTA:

fast iterative shrinkage/thresholding algorithm
 SALSA:

split augmented Lagrangian shrinkage algorithm
 WaTMRI:

wavelet tree sparsity MRI
 TVAL3:

total variation augmented Lagrangian alternating direction algorithm
 RecPF:

reconstruction from partial Fourier data
 TVCMRI:

total variation based compressed MRI
 FCSA:

fast composite splitting algorithm
 FADMM:

fast alternating direction method of multipliers
 ALPADMM:

accelerated linearized preconditioned alternating direction method of multipliers
 FLPADMM:

fast linearized preconditioned alternating direction method of multipliers
 SNR:

signaltonoise ratio
 Rel.Err:

relative error
 SSIM:

structural similarity index
References
 1.
Donoho DL. Compressed sensing. IEEE Trans Inf Theory. 2006;52(4):1289–306.
 2.
Lustig M, Donoho DL, Santos JM, Pauly JM. Compressed sensing MRI. IEEE Signal Process Mag. 2008;25(2):72–82.
 3.
Candès EJ, Wakin MB. An introduction to compressive sampling. IEEE Signal Process Mag. 2008;25(2):21–30.
 4.
Candès EJ, Romberg J, Tao T. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans Inf Theory. 2006;52(2):489–509.
 5.
Lustig M, Donoho D, Pauly JM. Sparse MRI: The application of compressed sensing for rapid MR imaging. Magn Reson Med. 2007;58(6):1182–95.
 6.
Afonso MV, BioucasDias JM, Figueiredo MA. Fast image recovery using variable splitting and constrained optimization. IEEE Trans Image Process. 2010;19(9):2345–56.
 7.
Eckstein J, Bertsekas DP. On the DouglasRachford splitting method and the proximal point algorithm for maximal monotone operators. Math Program. 1992;55(1–3):293–318.
 8.
Glowinski R, Le Tallec P. Augmented Lagrangian and operatorsplitting methods in nonlinear mechanics, vol. 9; 1989.
 9.
Goldstein T, O’Donoghue B, Setzer S, Baraniuk R. Fast alternating direction optimization methods. SIAM J Imaging Sci. 2014;7(3):1588–623.
 10.
Ouyang Y, Chen Y, Lan G, Pasiliao E Jr. An accelerated linearized alternating direction method of multipliers. SIAM J Imaging Sci. 2015;8(1):644–81.
 11.
Yang ZZ, Yang Z. Fast linearized alternating direction method of multipliers for the augmented\(\backslash\) ell _1regularized problem. Signal Image Video Process. 2015;9(7):1601–12.
 12.
Daubechies I, Defrise M, De Mol C. An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Commun Pure Appl Math. 2004;57(11):1413–57.
 13.
BioucasDias JM, Figueiredo MA. A new twist: twostep iterative shrinkage/thresholding algorithms for image restoration. IEEE Trans Image Process. 2007;16(12):2992–3004.
 14.
Beck A, Teboulle M. A fast iterative shrinkagethresholding algorithm for linear inverse problems. SIAM J Imaging Sci. 2009;2(1):183–202.
 15.
Chen C, Huang J. Exploiting the wavelet structure in compressed sensing MRI. Magn Reson Imaging. 2014;32(10):1377–89.
 16.
Li C. An efficient algorithm for total variation regularization with applications to the single pixel camera and compressive sensing. PhD thesis, Citeseer; 2009.
 17.
Ma S, Yin W, Zhang Y, Chakraborty A. An efficient algorithm for compressed MR imaging using total variation and wavelets. In: Computer vision and pattern recognition, 2008. IEEE conference on CVPR; 2008. p 1–8.
 18.
Yang J, Zhang Y, Yin W. A fast alternating direction method for tvl1l2 signal reconstruction from partial fourier data. IEEE J Sel Top Signal Process. 2010;4(2):288–97.
 19.
Huang J, Zhang S, Metaxas D. Efficient MR image reconstruction for compressed MR imaging. Med Image Anal. 2011;15(5):670–9.
 20.
Gabay D, Mercier B. A dual algorithm for the solution of nonlinear variational problems via finite element approximation. Comput Math Appl. 1976;2(1):17–40.
 21.
Monteiro RD, Svaiter BF. Iterationcomplexity of blockdecomposition algorithms and the alternating direction method of multipliers. SIAM J Optim. 2013;23(1):475–507.
 22.
Nesterov Y. A method for unconstrained convex minimization problem with the rate of convergence o (1/k2). Doklady an SSSR. 1983;269:543–7.
 23.
Nesterov Y. Introductory lectures on convex optimization: a basic course vol. 87; 2013.
 24.
Chang Y, King KF, Liang D, Wang Y, Ying L. A kernel approach to compressed sensing parallel MRI. In: 2012 9th IEEE international symposium on biomedical imaging (ISBI). IEEE; 2012. p. 78–81.
 25.
Liu B, Zou YM, Ying L. Sparsesense: application of compressed sensing in parallel MRI. In: 2008 international conference on information technology and applications in biomedicine. IEEE; 2008. p. 127–30.
 26.
King K. Combining compressed sensing and parallel imaging. In: Proceedings of the 16th annual meeting of ISMRM, Toronto, vol. 1488; 2008.
Authors' contributions
CSS conceived the study and participated in its design, implemented the algorithm, performed the acquisition of MR images, carried out the experiments, analyzed the data and drafted the manuscript. DHW participated in the design of the study, performed critical revision of the manuscript for important intellectual content and obtained funding. WLN also performed the acquisition of MR images and helped to draft the manuscript. JJQ drew some of the figures and supervised the study. QBS expressed opinions on the overall framework of the manuscript. All authors read and approved the final manuscript.
Acknowledgements
The authors would like to thank the Centers for Biomedical Engineering, University of Science and Technology of China for providing the 3T GE MR750 scanner. The authors are also grateful to the anonymous reviewers for their valuable comments and suggestions.
Competing interests
The authors declare that they have no competing interests.
Availability of data and materials
All datasets related to the current study are available from the corresponding author on reasonable request.
Consent for publication
In this study, all of the experiments were agreed upon by the volunteers and Centers for Biomedical Engineering, University of Science and Technology of China, Heifei, China. In addition, all MR datasets would be published.
Ethics approval and consent to participate
The ethical approval was given by the Medical Ethics Committee of the University of Science and Technology of China, Heifei, China.
Funding
This work was supported by the National Natural Science Foundation of China under Grant No .81527802.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Author information
Affiliations
Corresponding author
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Chen, S., Du, H., Wu, L. et al. Compressed sensing MRI via fast linearized preconditioned alternating direction method of multipliers. BioMed Eng OnLine 16, 53 (2017). https://doi.org/10.1186/s129380170343x
Received:
Accepted:
Published:
Keywords
 Compressed sensing MRI
 Image reconstruction
 Total variation
 Alternating direction method of multipliers