Journal of Signal and Information Processing, 20 11 , 2, 44 - 51
doi:10.4236/jsip.2011.21007 Published Online February 2011 (http://www.SciRP.org/journal/jsip)
Copyright © 2011 SciRes. JSIP
TV Sparsifying MR Image Reconstruction in
Compressive Sensing
Yonggui Zhu, Xiaolan Yang
School of Science, Communication University of China, Beijing, China
Email: ygzhu@cuc.edu.cn
Received January 23rd, 2011; revised February 18th, 2011; accepted February 19th, 2011
ABSTRACT
In this paper, we apply alternating minimization method to sparse image reconstruction in compressed sensing. This
approach can exactly reconstruct the MR image from under-sampled k-space data, i.e., the partial Fourier data. The
convergence analysis of the fast method is also given. Some MR images are employed to test in the numerical experi-
ments, and the results demonstrate that our method is very efficient in MRI reconstruction.
Keywords: Compressed Sensing, Magnetic Resonance Image, Total Variation, Image Reconstru ct ion
1. Introduction
Compressed Sensing has enormous potentials for scan
time reducing significantly in magnetic resonance im-
age(MRI) research community. Compressed Sensing
theory was put forward by Candes, Romberg and Tao
[1], and D. Donoho [2] in 2006. They pointed out that
sparse signals can be reconstructed from a very limited
number of samples, provided that the measurements
satisfy an incoherence property [3]. For MRI Com-
pressed Sensing, it is possible to reconstruct accurately
MR images from under-sampled k-space data, i.e.,
the partial Fourier data by solving optimization prob-
lems.
Suppose
N
uR is a sparse signal, and 0
u
denotes number of n on-zer os in u. Let A be
M
N
K <
M
N measurement matrix such that
A
ub, where
b is an observed data vector. Then to recover u from b is
equivalent to solve 0
Lproblem:

0
min ,
uuAub (1)
However, problem (1) is provably NP hard [4] and
very difficult to solve from the viewpoint of numerical
computation. Thus, i t is rexali stic t o solve 1
L problem:

1
min ,
uuAub (2)
which has also been known to yield sparse solutions un-
der some conditions (see [5,6] for explains). In the case
of Compressed Sensing MRI, A is a partial Fourier ma-
trix, i.e., ,
M
N
APFPR

is consisted of
M
N
rows of the identity matrix, F is a discrete Fourier matrix.
When b is contaminated with noise such as Gaussian
noise of variance 2
, the relaxation form for problem (2)
should be gi ven by
22
12
min :
uuAub
 (3)
The total variation regularization has been first pro-
posed for image denoising by Rudin, Osher and Fatemi
[7]. It is well known that TV regularizier can better re-
cover piecewise smooth signals with preserving sharp
edges or boundaries. So TV regularizer is a sparsifying
transform operator for piecewise smooth MR images
such as brain images. When only TV sparsifying trans-
form is considered, the optimization problem in Com-
pressed Sensing MRI can be written as
22
2
min ,
TV
uuPFu b
 (4)
The unconstrained version o f pro blem (4) is
2
2
min 2
TV
uuPFub
, (5)
where μ is a positive parameter that determines the
trade-off between the fidelity term and the sparsity term,
2
.denotes the Euclidean norm. Model (5) was previ-
ously mentioned by He etal.[8] and Lustig et al. [9].
In this paper, we focus on the two dimensional MRI
Compressed Sensing model (5). All two dimensional
images are changed into one dimensional vectors in the
context of paper. For square images, let 2
n
uR be a
TV Sparsifying MR Image Reconstruction in Compressive Sensing
Copyright © 2011 SciRes. JSIP
45
nn image. TV
u in (5) is defined as

1
22
2
1, 2,
,jk jk
jk
TV
uuu
. In the two dimen-
sional image practical applications, we only consider the
periodic boundary condition of the image. Therefore, the
discrete gradient operators 1
and 2
of nn
im-
age can be defined by
1, ,
1, 1, ,
jk jk
jk knk
uuifjn
uuu ifjn



(6)
,1 ,
2, ,1 ,
jk jk
jk jjn
uuifkn
uuu ifkn



(7)
 
22
12
,nn
DDR
denote finite difference matrices
corresponding to 1,
j
k
u and 2,
j
k
u respectively. Let




12 2
,T
iii
DuD uDuR where


,1,2
l
i
Du l
are the

ijnk th entry of

l
Du, i is correspond-
ing to the (j,k)th pixel position of the image. The summa-
tion 2
12
n
i
iDu
equals to the discrete total variation of u,
i.e. TV
uThus (5) is changed into
22
22
1
min 2
n
i
ui
DuPFu b

(8)
Recently, Y. L. Wang et al. [10] have studied the total
variation image restoration model:
2
2
min 2
TV
uuBuf
 (9)
i.e.
22
22
1
min 2
n
i
ui
DuBuf

, (10)
where 22
nn
BR
represents a blurring operator. They
proposed an alternating minimization method to solve (9).
In this paper, we will apply this approach to the Com-
pressed Sensing MRI model (8).
In model (8), 2
mn
PR
is a selection matrix con-
sisting of 2
mn rows of the identity matrix, F is a
two-dimensional discrete Fourier transform matrix. So
PF serves as a sensing matrix. b is acquired (by coils in
an MRI scanner) and sent to a computer, so b is an ob-
served data containing noise. We will employ quadratic
penalty approach [11] to split the model (8) into two
sub-minimization problems, and build the fast method to
solve the two sub-problems. We will also analyze the
convergence of the fast reconstruction method. This pa-
per is organized as follows. In section 2, basic algorithm
and optimization is given. Section 3 presents the conver-
gence analysis of the fast reconstruction algorithm.
In section 4, Shepp-Logan Phantom image and some
real MR images are employed to do numerical experi-
ments to demonstrate the effectiveness of the fast recon-
struction method.
2. Basic Algorithm and Optimization
Firstly, we introduce some notations for the sake of con-
venience. For vectors 2
n
iR
and finite difference ma-
trices

22
,1,2,
inn
DR i
 assume


1212
;,
T
TT
w
 

and
 
 
12 12
;,
T
TT
DDD DD
 Let
2
22
12
();(), 1,2,,
iii
n
wRin
vR



By using quadratic penalty approach, we can change
(8) into the following prob lem:
22
22
222
,, 11
min 22
nn
iii
uw ii
wwDuPFub







 (11)
In (11), i
w is the approximation of i
Du. It is well
known that the solution of (11) converges to that of (8) as
. Therefore the solution of (11) for
large
enough can better approximate to that of (8). If fix alter-
natively u and w, we can obtain the following two
sub-problems:
22
2
2
11
2
min 2
nn
iii
wii
wwDu


 , (12)
222
2
12
min 22
n
ii
ui
wDu PFub

 
. (13)
Since w-subproblem (12) is separable with respect to
i
w, minimizing (12) is equivalent to solving
22
22
min,1,2, ,
2
iiii
wwin
wDu

(14)
Therefore, we can get its minimizer by two dimen-
sional shrinkage:
2
,1,2,,
iwi
wsDuin, (15)
where

1
max,0 i
wii i
Du
sDuDu Du




in which
0
00
0


 is assumed.
For vectors 2
,n
x
yR, define

22
22
,:nn
w
SxyR R
by
 

2
12
,,,,
T
wwww
n
Sxy szszsz, (16)
where

2
,,1,2,,
T
iii
zxyi n
TV Sparsifying MR Image Reconstruction in Compressive Sensing
Copyright © 2011 SciRes. JSIP
46
Rewrite (15) as the following:


12
,
w
wsDuDu. (17)
For u-subproblem (13), the first-order optimality con-
dition is
 
22
11 0
nn T
TT
ii ii
ii
DDuDwPFPFu b


 
 (18)
We note that

2
(1)(1) (2)(2)
1
TT
nTT
ii
i
DDuDDDDuDDu
 
2
1
nTT
ii
i
Dw Dw
Thus

TTT TTT
DDFPPFuDw FPb
 

. (19)
For l = 1, 2, suppose the Fourier transform of

l
D are

l
D
. Let
 

 
12 12
;,
T
TT
DDD DD

 
Apply the Fourier transform to both sides of (19), we
can obtain
TT TT
DDFuPPFuDFw Pb


 . (20)
For the periodic boundary condition for u, (1)(1) ,
T
DD
(2) (2)
T
DD are all block circulant [12]. Therefore, T
DD
can be diagonalized by Fourier transform. It is easy to
see that T
PP is diagonal. Thus TT
DD PP
 is also
a diagonal matrix. Given w, we can obtain u by the two
steps. The first step is to get
F
u by solving (20). The
second step is to obtain u by applying the inverse fou-
rier transform to
F
u. For the sake of simplicity, let

222
2
2
1
min 22
n
uii
ui
s
wwDuPFub


. (21)
For a fixed
, the alternating approximation algo-
rithm for solving (8) can be given as follows:
Algorithm 1 Input b, Pand parameters
> 0, μ
> 0. Initialize

0
uu
While stopping criterion is not satisfied, Do
Compute
 

 

1121
,,
,1,2,.
kkk
w
kk
u
wsDuDu
uswk


End Do
To solve the u-subproblem use only a FFT transform
and a inverse FFT transform, and the solution of w-
subproblem can be obtained by two dimensional shrink-
age, so this alternating algorithm is a fast approach. The
stopping criterion will be given in the section numerical
experiments. The convergence analysis of this algorithm
will be discussed in detail in th e next section.
3. The Convergence Analysis and a
Continuation Strategy
In this section, we give the convergence analysis of the
alternating algorithm for fixed
> 0. We give the defi-
nition of the firmly non-expansive operator before the
presentation of non-expansiveness for two dimensional
shrinkage.
Definition 3.1. An operator 22
:
s
RR is firmly non-
expansive if it satisfies the following condition:
 
22
2
2
22
s
x syxyIdsxIdsy
where Id is identity operator. It is easy to show that a
firmly non-expansive operator22
:
s
RR is non-ex-
pansive, i.e., for any
 
2
2
2
,,
x
yRsx syxy
Lemma 3.1. If
22
:PR R is the projection onto
21
xRx
 

, and for 2
x
R the two dimen-
sional shrinkage operator 22
:
w
s
RR is defined as
22
1
max ,0
w
x
sx
x



where0
00
0



 is fol-
lowed, then
w
s
xxPx .
Lemma 3.2. For all 2
,
x
yR, it holds that
 
22
2
2
22
ww
s
xsyxy PxPy Further-
more, if
2
2
ww
s
xsyxy
 ,.then

2
2
22
,:nn
w
S
xy RR
x
y
The proofs of these two lemmas are shown in [10].
By Lemma 3.1 and Lemma 3.2, we can get the fol-
lowing theorem:
Theorem 3.1. For 2
x
R, if the two dimensional
shrinka g e operator 22
:
w
s
RR is given by
22
1
max ,0
w
x
sx
x



then

w
sis non-expansive,
i.e., for a n y 2
,
x
yR,
2
2
ww
s
xsyxy. (22)
Theorem 3.2. For vectors 2
,n
x
yR, if
22
22
,:nn
w
SxyR R is defined by
 

2
12
,,,,
T
wwww
n
Sxy szszsz
where

2
,,1,2,,
T
iii
z
xy inand
 

12
;DDD
 
12
,
TT
T
DD , then
TV Sparsifying MR Image Reconstruction in Compressive Sensing
Copyright © 2011 SciRes. JSIP
47





12 12
2
2
,,
ww
SDxDxSDyDyDxy(23)
Proof. By theorem 3.1 and the definition of w
S, for
any 2
,n
x
yR, we obtain




















2
2
2
12 12
2
12
1
2
12
2
2
12 12
12
2
2
,,
,
,
,,
ww
nT
wii
i
T
wii
nTT
ii ii
i
SDxDxSDyDy
sDxDx
sDyDy
Dx DxDy Dy
Dx Dy








since
 

 

12 12
;,
TT
T
DDD DD

This completes the proof.
Let TTT
M
DD FPPF

 , since 0, 0

, M
is nonsingular. According to (19), we have
 



 


  

1
1121
1
1121
11
,
,
kk
u
k
TTT
kk
TTT
w
kk
TTT
w
uSw
MDw FPb
M
DSD uDuFPb
M
DSD uDuMFPb









Assume


 

11121
1
1
,
kkk
Tw
TT
uMDSDuDu
MFPb


, (24)
Then
 

1kk
uu
 (25)
Next we show that the operator T is non-expansive.
Theorem 3.3. For μ large enough in (8), the operator
T in (24) is non-expansive, i.e., for 2
,n
x
yR, it holds
that
 
2
2
TxTyx y
Proof.
 
 




 

 


2
12 12
1
2
12 12
1
2
,,
,,
Tww
Tww
Tx Ty
MDS DxDxS DyDy
MDS DxDxS DyDy


.
Using (23) in Theorem 3.2, we have
 
 


12 12
1
2
1
2
,,
Tww
T
MDS DxDxS DyDy
MDDxy

Hence
 

1
22
1
2
1
2
T
T
T
TxTyM DDxy
MDDxy
DDMx y
 


where
is matrix-norm.
Since TTT
M
DD FPPF

 , if μ is large enough
such that
11
T
DDM
then

2
Tx Ty
2
x
y
This completes the non-expansiveness of the operator
T (·).
Theorem 3.4. For any initial value

2
0n
u, as-
sume

k
u be generated by (25), then T is asymptoti-
cally regular, i.e.,
 



100
1
22
lim lim0
kk kk
kk
uu TuTu
 

Proof. Since
 
1121
,
kkk
w
wSDuDu

by (19), we have



1
12
,
k
TTT
kk
TTT
w
DD FPPFu
DSD uDuFPb



(26)

 

1121
,
k
TTT
kk
TTT
w
DD FPPFu
DSD uDuFPb




. (27)
(26) minus (27) is
 
  

 


1
12 1121
,,
kk
TTT
kkk k
Tww
DDFPPF uu
DSDuDuSDuDu




That is
 
 

 

1
12
1
1121
,
,
kk
kk
Tw
kk
w
uu
MDSDuDu
SDu Du

Therefore
 


 


11
2
12 1121
2
,,
kk T
kkk k
ww
uu MD
SDu DuSDuDu


TV Sparsifying MR Image Reconstruction in Compressive Sensing
Copyright © 2011 SciRes. JSIP
48
Using non-expansiveness of w
S, we obtain
  

11
1
22
kk kk
T
uu MDDuu


That is
 
11
22
kk kk
uu uu

 
where 1T
M
DD

When μ is large enough such that 1
, then we have
 
110
22
kk k
uu uu
 
So
 
1
2
lim 0
kk
kuu
 
This completes the proof.
We note that the objective function in (11) is convex,
bounded below, and coercive, thus (11) has at least one
minimizer

,uw

, and must satisfy
u
uSw

 

12
,
w
wSDuDu

So u is a fixed point of T.
According to the Opial theorem [13], the squence


k
u converges to a fixed point of T.
The algorithm 1 given in the section 2 can be signifi-
cantly accelerated by employing a continuation scheme
introduced by Y. Wang [10]. That means penalty pa-
rameters
vary with k, starting from initial small val-
ues and increase them gradually. In the implementation
of the algorithm, firstly for small fixed
apply the
algorithm to solve (11) up to the stopping criterion. Next
the obtained solution is used as new starting point of ap-
plying the algorithm to (11) for the next
. Go on like
this until the the given max
is obtained. Such a con-
tinuation scheme is also called path-following technique
which is widely used in the penalty methods [14-17].
This accelerating convergence result is also verified by
our numerical experiments. Now we present the fast re-
construction method (FRM) for MR images with TV
sparsity, which will be be used in the numerical experi-
ments.
Algorithm 2 Input b, P, μ > 0, 00
and
max 0

. Initialize

0
uu
While max

, Do
1) For fixed
, run Algorithm 1 until the stopping
criterion is satisfied.
2) Update 2
 .
End Do
4. Numerical Experiments
In this section, we present the performance of fast recon-
struction method(FRM) for MR images with TV sparsity.
We compare our method with Two-step iterative shrink-
age/ thresholding algorithm(TwIST) [18], which can be
used to solve (8). The general model solved by TwIST is

2
2
1
min 2
u
A
ub u
, (28)
where
is a reg ularization functio n such that the so lu-
tion of the den ois i n g pr obl em
 
2
2
1
min 2
u
vvuu
 , (29)
is known. If set

2
12
1
,,
n
i
i
uDu APF
 
, then
(28) is identical to (8). That is the reason that we choose
TwIST method to compare with our method.
Here we give the iteration framework of TwIST to
solve (28) in brief. TwIST is a two-step version of itera-
tive shrinkage/thresholding(IST) [19] algorithm. The
iteration formula of TwIST is as follows:


2
2
1
min 2
kk
u
vuvu
, (30)

11
1
kk kk
uu uv


 , (31)
where ,0

T
kk k
vuAbAu  , In the latest
software, TwIST_v1,
and
are set as
2
ˆˆ1

, (32)

1
ˆ
ˆ2
m

 . (33)
In which, the detail explain of parameters 1
, m
and
ˆ
can be seen in [18]. In TwIST_ v1, denoising prob-
lem (30) is solved iteratively by Chambolle’s algorithm
[20]. Signal to noise ratio (SNR) and relative error
(ReErr) are used to measure the quality of the recon-
structed images. The definitions of SNR and ReErr are
given as follows:
02
02
10log10 u
SNR uu




, (34)
2
02
2
02
Re uu
Err u
, (35)
where u and 0
u are the reconstructed image and
original image, respectively. All experiments were done
in MATLAB on a laptop with Intel Core Duo P8400
processor and 2 GB of memory.
The 256 × 256 Shepp-Logan phantom image is sam-
pled with 22 views in frequency space, for which there
are 6136 Fourier coefficients measured, i.e., sample ratio
is 9.36%. Figure 1 shows 22 radial lines in frequency
space.
TV Sparsifying MR Image Reconstruction in Compressive Sensing
Copyright © 2011 SciRes. JSIP
49
Figure 1. Sampling domain with 22 views in frequen cy sp ace.
In the following tests, we assume the mean value and
standard deviation for the additive Gaussian noise are 0
and 0.01, respectively. When using FRMto do numerical
experiments, we initialize 5
2
and final max
equals
to 10
2. Let μ = 1000, and set initial image

0
u to zero.
The stopping criterion is the relative difference between
the successive iterate of the reconstructed image satisfy
the following:
 

1
4
2
2
10
kk
k
uu
u
(36)
According to the above mentioned parameters and the
stopping criterion, we use our algorithm to do Phantom
image reconstruction experiment. Figure 2(a) is Phepp-
Logan phantom original image. The reconstructed image
is showed in Figure 2(b). The SNR, ReErr and CPU
time(seconds) are 31.3687 dB, 0.0270 and 19.4375, re-
spectively.
Now we apply TwIST method to do Phantom image
reconstruction experiment. We choose the monotonic
variant of TwIST in TwIST_v1, which stops when the
relative change in the objective function falls below Tol-
erance=1e-4. Parameters
and
are determined by
using (32) and (33). The parameter 1
is equal to 1e-3
that is recommended by the TwIST_v1 documentation.
Set
= 1e-3, which is corresponding to μ = 1000.
If we use TwIST to do the reconstruction test for the
Logan Phantom image with same sampling ratios. The
result is presented in Figure 2(c). Its SNR, ReErr and
CPU time(seconds) are 29.0010 dB, 0.0355 and 55.5938,
respectively. From all the two reconstructed images, it is
clear that our method have better qualities than TwIST.
In addition, from the values of SNR, ReErr and CPU
time for restructed image, our method increases 11 the
SNR by 2.3677 dB and reduce the ReErr and CPU time
0.0085 and 35.1563 seconds more than TwIST.
When we choose some views 44 , 66, 88, the sampling
ratios for Phantom image are 18.76%, 26.85%, 34.97%.
The reconstructed results are showed in Figure 3 and
Figuure 4 by the two methods.
The following table presents the values of SNR, ReErr
and CPU time of reconstructions under different views
by FRM and TwIST.
From Table 1, we can see that the SNR obtain ed using
FRM is higher than that obtained using TwIST under the
same sampling ratios or views. Figure 5(a), (b) and (c)
give SNRs, the relative errors and CPU time, respec-
tively, of reconstructed images by FRM (red curves) and
TwIST (green curves) from the observed data at a se-
quence of different sampling ratios.
Now we look three different 256 × 256 MR brain im-
ages, which are shown in Figure 6 (a), (b) and (c), re-
spectively.
We choose 66 views sampling for the three MR im-
ages, in which the sampling ratio is equal to 26.85%.
(a) (b) (c)
Figure 2. (a) Original image. (b) Reconstruction by FRM. (c)
Reconstruction by TwIST.
(a) (b) (c)
Figure 3. Reconstruction by FRM (a) 44 views. (b) 66 views.
(c) 88 views.
(a) (b) (c)
Figure 4. Reconstruction by TwIST (a) 44 views. (b) 66 views.
(c) 88 views.
TV Sparsifying MR Image Reconstruction in Compressive Sensing
Copyright © 2011 SciRes. JSIP
50
Table 1. SNR, ReErr, CPU of reconstructions under differ-
ent views or sampling ratios by two methods.
Views/Samp.ratio Method SNR ReErr CPU
FRM 40.6877 0.0092 7.1875
44/18.76% TwIST 30.7696 0.0289 26.4219
FRM 44.8714 0.0057 4.9063
66/26.85% TwIST 30.8412 0.0287 20.2031
FRM 47.8810 0.0040 3.6719
88/34.97% TwIST 30.8412 0.0286 17.7969
(a)
(b)
(c)
Figure 5. (a) SNRs. (b) Relative Errors. (c) CPU time.
When use FRM and TwIST to do reconstruction ex-
periments for the three MR images, all parameters in the
two methods are still same as the above mentioned. The
results reconstructed by FRM and TwIST are presented
in Figure 7. and Figure 8., respectively.
The SNRs, the relative errors and CPU time of the re-
constructed images by FRM and TwIST are given in the
Table 2.
From the quality of reconstructed images and the val-
ues of SNR, ReErr and CPU time, we can see that FRM
method is better than TwIST method.
(a) (b) (c)
Figure 6. Original image. (a) Brain 1. (b) Brain 2. (c) Brain 3.
(a) (b) (c)
Figure 7. Reconstructions by FRM (a) Brain 1. (b) Brain 2.
(c) Brain 3.
(a) (b) (c)
Figure 8. Reconstructions by TwIST (a) Brain 1. ( b) Br ain 2.
(c) Brain 3.
Table 2. SNR, ReErr, CPU time and sampling ratios for
reconstructed results by FMR and TwIST.
ImageMethodSNR ReErr CPU Samp.ratio
FRM 26.12880.0494 6.9063 26.85%
Brain
1 TwIST 24.73650.0580 20.8438 26.85%
FRM 21.93530.0800 9.1719 26.85%
Brain
2 TwIST 21.14940.0876 20.9375 26.85%
FRM 27.69160.0412 6.4531 26.85%
Brain
3 TwIST 26.15990.0492 20.5938 26.85%
TV Sparsifying MR Image Reconstruction in Compressive Sensing
Copyright © 2011 SciRes. JSIP
51
5. Conclusion
We have developed a fast reconstruction method for
compressed sensing MRI that is called FRM. The con-
vergence of the reconstruction method has been analyzed.
The algorithm has been accelerated by continuation on
penalty parameters. FRM is compared with TwIST, a
state-of-the-art method. We use the two methods to re-
construct Phantom image and some real MR images from
a partial of observed data. The results of numerical ex-
periments on the MR images demonstrate that FRM can
achieve much higher performance in terms of SNRs, the
relative errors and CPU time than TwIST. We believe
that our method can be applied in the area of rapid MR
imaging.
6. Acknowledgements
This work was supported by the Key Project of Chinese
Ministry of Education(No. 109030), 382 Training Pro-
gramme of CUC(G08382316) and Science Research Pro-
ject of CUC(XNL1003). The authors would like to thank
the referees for very useful suggestions here.
REFERENCES
[1] E. J. Candes, J. Romberg, and T. Tao, “Robust uncer-
tainty principles: Exact signal reconstruction from highly
incomplete frequency information,” IEEE Trans. Inf. The-
ory, Vol. 52, 2006, pp. 489-509.
[2] D. L. Donoho, “Compressed sensing,” IEEE Trans. Inf.
Theory, Vol. 52, 2006, pp. 1289-1306.
[3] E. Candes and J. Romberg, “Sparsity and incoherence in
compressive sampling,” Inverse Problems, Vol. 23, 2007,
pp. 969-985.
[4] B. K. Natarajan, “Sparse approximate solutions to linear
systems,” SIAM J. Comput., Vol. 24, 1995, pp.227-234.
[5] D. Donoho and X. Huo, “Uncertainty principles and ideal
atomic decompositions,” IEEE Trans. Inf.Theory, Vol. 47,
2001, pp. 2845-2862.
[6] J. J. Fuchs, “On sparse representations in arbitrary redun-
dant bases,” IEEE Trans. Inf. Theory, Vol. 50, 2004, pp.
1341-1344.
[7] L. Rudin, S. Osher and E. Fatemi, “Nonlinear total varia-
tion based noise removal algorithms,” Physica D, Vol. 60,
1992, pp. 259-268.
[8] L. He, T. C. Chang, S. Osher, T. Fang and P. Speier, “MR
image reconstruction by using the iterative refinement
method and nonlinear inverse scale space methods,”
UCLA CAM Report, 2006,06-35.
[9] M. Lustig, D. Donoho and J. Pauly, “Sparse MRI: The
application of compressed sensing for rapid MR imag-
ing,” Magnetic Resonance in Medicine, Vol. 58,2007,pp.
1182-1195.
[10] Y. Wang, J. Yang, W. Yin and Y. Zhang, “A alternating
minimization algorithm for total variation image recon-
struction,” SIAM J. Imag. Sci., Vol. 1, 2008,pp. 248-272.
[11] R. Courant, “Variational methods for the solution of
problems with equilibrium and vibration,” Bull.Amer.
Math. Soc., Vol. 49, 1943, pp. 1-23.
[12] M. K. Ng, R. H. Chan and W. C. Tang, “A fast algorithm
for deblurring models with Neumann boundary condi-
tions,” SIAM J. Sci. Comput., Vol. 21, pp. 851-866, 1999.
[13] Z. Opal, “Weak convergence of the sequence of succes-
sive approximations for nonexpansive mappings,” Bull.
Amer. Math. Soc., Vol. 73, 1967, pp. 591-597.
[14] J. Yang, W. Yin and Y.Wang, “A Fast Algorithm for
Edge-Preserving Variational Multichannel Image Resto-
ration,” SIAM J. Sci. Comput., Vol. 2, 2009, pp. 569-592.
[15] J. Yang, Y. Zhang and W. Yin, “A Fast Alternating Di-
rection Method for TVL1-L2 Signal Reconstruction From
Partial Fourier Data,” IEEE Journal of Selected Topics in
Signal Processing, Vol. 4, 2010, pp.288-297.
[16] E. T. Hale, W. Yin and Y. Zhang, “A Fixed-Point Con-
tinuation for l1regularization with Application to Com-
pressed Sensing,” Rice University CAAM Technical Re-
port TR07-07, 2007, pp. 1-45.
[17] E. T. Hale, W. Yin and Y. Zhang, “Fixed-Point Continua-
tion for l1Minimization: Methodology and Conver-
gence,” SIAM J. Sci. Comput., Vol. 2, 2009, pp. 569-592.
[18] J. Bioucas-Dias and M. Figueiredo, “A new TwIST:
Two-step iterative thresholding algorithm for image res-
toration,” IEEE Trans. Imag. Process., Vol. 16, 2007, pp.
2992-3004.
[19] I. Daubechies, M. Defriese and C. De Mol, “An iterative
thresholding algorithm for linear inverse problems with a
sparsity constraint,” Commun. Pure Appl. Math., Vol. 57,
2004, pp. 1413-1457.
[20] A. Chambolle, “An algorithm for total variation minimi-
zation and applications,” J. Math. Imaging Vision, Vol.
20, 2004, pp. 89-97.