Transactions of Nanjing University of Aeronautics & Astronautics
网刊加载中。。。

使用Chrome浏览器效果最佳,继续浏览,你可能不会看到最佳的展示效果,

确定继续浏览么?

复制成功,请在其他浏览器进行阅读

A Regularized Randomized Kaczmarz Algorithm for Large Discrete Ill‑Posed Problems  PDF

  • LIU Fengming 1
  • WANG Zhengsheng 1
  • YANG Siyu 1
  • XU Guili 2
1. College of Science, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, P. R. China; 2. College of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, P. R. China

CLC: O241.6

Updated:2020-11-23

DOI:http://dx.doi.org/10.16356/j.1005-1120.2020.05.013

  • Full Text
  • Figs & Tabs
  • References
  • Authors
  • About
OUTLINE

Abstract

Tikhonov regularization is a powerful tool for solving linear discrete ill‑posed problems. However, effective methods for dealing with large‑scale ill‑posed problems are still lacking. The Kaczmarz method is an effective iterative projection algorithm for solving large linear equations due to its simplicity. We propose a regularized randomized extended Kaczmarz (RREK)algorithm for solving large discrete ill‑posed problems via combining the Tikhonov regularization and the randomized Kaczmarz method. The convergence of the algorithm is proved. Numerical experiments illustrate that the proposed algorithm has higher accuracy and better image restoration quality compared with the existing randomized extended Kaczmarz (REK) method.

0 Introduction

This paper mainly considers solving the large discrete ill‑posed problem

1

minxRnAx-b̃2 (1)

where ARm×n is an ill‑conditioned matrix and its singular value is gradually reduced to zero without obvious intervals, the vector b̃Rm is error‑contaminated data, that is

b̃=btrue+r,btrueRA,rRA

where btrue is the output result in the ideal state, and r the unavoidable noise data during the observation process. Assume that the linear equation

Ax=btrue (2)

is consistent, we will determine its solution by calculating the approximate solution of the linear discrete ill‑posed problem (1).

In practical applications, Tikhonov regularization

2 is one of the most commonly used methods for solving linear discrete ill‑posed problems. For small‑scale problems, the solution of the ill‑posed problem can be obtained by selecting appropriate regularization parameters with the direct regularization method. However, for large‑scale ill‑posed problems, the direct application of Tikhonov regularization method needs a large amount of computation and storage.

Kaczmarz method

3is an effective iterative projection method for solving large-scale consistent linear Eq.(2). Due to its simplicity, it has been widely used in image reconstruction, signal processing, distributed computation and other fields4-6. In order to improve the convergence of the Kaczmarz method, Strohmer and Vershyin7proposed a randomized Kaczmarz (RK) method with exponential convergence, which randomly selects the rows of the matrix A. For the inconsistent problem, Needell8proved that the randomized Kaczmaz method could converge to the neighborhood of the least squares solution, and the radius of the neighborhood were related to the noise of the inconsistent problem. Inspired by Popa method9, Zouzias and Freris10put forward a randomized extended Kaczmarz (REK) method, which makes the inconsistent problems converge to the least squares solution.

In consequence, for large-scale ill-posed problems, this paper considers combining the REK method with Tikhonov regularization, so as to generate a regularized iterative method.

The structure of the paper is as follows: Section 1 introduces the Kaczmarz method and proposes the regularized randomized extended Kaczmarz (RREK)algorithm for the ill-posed problem. In Section 2, the convergence of the new algorithm is proved. In Section 3, we carry out some numerical experiments. Finally, the relevant conclusions are drawn.

1 Regularized Randomized Extended Kaczmarz Algorithm

1.1 Kaczmarz method

The classical Kaczmarz algorithm is an iterative projection method for solving large linear consistent equations, the algorithm starts with an arbitrary vector x0. In each iteration, the rows of the matrix are traversed in a circular manner. For each selected row, the current iteration point xk-1 is orthogonally projected onto the next hyperplane Hi:={x |<Ai,x>=bi}, and the projection point is used as the next iteration point xk, the resulting sequence converges to the solution of Eq.(2

11. Given the initial value, the iterative formula of Kaczmarz algorithm is as follows

xk=xk-1-<xk-1,Ai>-biAi22Ai
k=0,1,2, (3)

where i=kmodm+1, 〈,〉 is the Euclidean inner product, Ai the transpose of the ith row vector of A,and bi the ith element of b.

From Eq.(3), we can see that the Kaczmarz method is very simple since it mainly contains the inner product operation, but its convergence rate is usually very slow. In order to improve the convergence of the Kaczmarz method, Strohmer and Vershyin proposed to randomly select the rows of the matrix A according to a probability proportional to Ai22/AF2i=1,2,,m. This method is called the RK method

7 and has exponential convergence.

For inconsistent problems, the above algorithm is no longer applicable. Zouzias and Freris proposed the REK method, which was a combination of randomized orthogonal projection algorithm and randomized Kaczmarz algorithm. REK algorithm

1012 was mainly used to solve the ill-posed problem (1), and described as follows:

The main idea of Algorithm 1 is to eliminate the noise part of the right-hand term by randomized orthogonal projection, and then apply the randomized Kaczmarz algorithm to the new linear system. The right-hand vector of the linear system is now arbitrarily close to the column space of A, i.e.Axbtrue, which makes the inconsistent problem converge to the least squares solution.

Algorithm 1 REK

1. Input: ARm×n, bRm, ε

2. Initial: x0Rn,y0=b̃

3. for k=1,2,3,

4. Randomly select jkn, compute

yk=yk-1-<yk-1,Ajk>Ajk22Ajk

5. Update bk,bk=b̃-yk

6. Randomly select ikm, compute

xk=xk-1-<xk-1,Aik>-bikkAik22Aik

7. Terminate if it holds xk-xk-12xk2<ε

8. end for

1.2 Randomized extended Kaczmarz algorithm based on Tikhonov regularization

Due to the ill-posedness property of problem (1), it is usually necessary to regularize the original problem. Tikhonov method is the most commonly used regularization method to solve ill-posed problems by replacing the minimization problem (1) with the penalized least squares problem

minxRnAx-b̃22+ωLx22 (4)

The regularization term ωLx22 is used to control the smoothness or sharpness of the solution, where ω is the regularization parameter and L the discrete approximation of some derivative operators.

Solving problem (4) is equivalent to solve

minxRn    AωLx-b̃022 (5)

The system of normal equations for the problem can be written as

AT,ωLTb̃0-AωLx=0

Let r1r2=b̃0-AωLx, then

AωLx=b̃0-r1r2AT,ωLTr1r2=0 (6)

Owing to the limitation of the storage and computation, direct regularization method is often used to solve small-scale ill-posed problems. However, for some inverse mathematical physics problems, the order of the discretized coefficient matrix may be very large. Therefore, based on Tikhonov regularization, this paper considers combining Kaczmarz method to solve large-scale linear discrete ill-posed problems.

In this paper, we use the Morozov discrepancy principle to determine the value of the regularization parameter ω and L is selected as the first derivative operator.

Using Eq.(6) and the idea of Algorithm 1, we can get Algorithm 2.

Algorithm 2 RREK

1. Input:ARm×n,LRn×n is the first derivative operator,b̃Rm,ω,ε; let

A¯=    AωL,b¯=b̃0,r=r1r2,b=b¯-r

2. Initial:x0=0, y0=b¯

3. for k=1,2,3,

4. Pick jkn with probability

qj:=A¯j22/A¯F2

5. Compute

yk=Im-A¯jkA¯jkTA¯jk22yk-1

6. Take the first m lines of yk, denoted as y1k, and take the last n lines of yk, denoted as y2k, then b1k=b̃-y1k, b2k=-y2k.

7. Pick ikm with probability

pi:=A¯i22/A¯F2

8. If 1ikm, compute

xk=xk-1-<xk-1,A¯ik>-b1kikA¯ik22A¯ik

if m+1ikm+n, let ik:=ik-m, compute

xikk=xikk-1+ω2(xik+1k-1-xikk-1)-b2kikω
xik+1k=xikk-1-ω2(xik+1k-1-xikk-1)-b2kik+1ω

9. Terminate if it holds

A¯xk-(b¯-yk)2A¯Fxk2ε,  A¯Tyk2A¯F2xk2ε

10. end for

2 Convergence Analysis

We describe a random algorithm (Algorithm 2), which consists of two parts. The first part, consisting of Steps 4 and 5, applies randomized orthogonal projection algorithm to maintain an approximation to b formed by b¯-yk. The second part is the randomized Kaczmarz algorithm composed of Steps 7 and 8. This paper first proves that b¯-ykb in the randomized orthogonal projection algorithm, and then proves the linear convergence of Algorithm 2.

Lemma 1  

10     For every vector uR(A¯), it holds

Im-A¯A¯TA¯F2u21-σmin2A¯F2u22

where σmin is the minimum singular value of A¯.

Theorem 1   For any matrix A¯, right‑hand side vector b¯, after k iterations, the randomized orthogonal projection algorithm has the expected convergence rate

Eyk-r221-1κF2(A¯)kb22

where κF(A¯)=A¯FA¯+2A¯+is the Moore-Penrose pseudo-inverse of A¯.

Proof   Define P(j):=Im-A¯jA¯jT/A¯j22 for every jn. Notice that P(j)P(j)=P(j), i.e.P(j) is a projection matrix.

Let X be a random variable and choose the index j according to the probability A¯j22/A¯F2, obviously E[PX]=Im-A¯A¯T/A¯F2.

For every k, define ek:=yk-r, it holds that

ek=P(jk)ek-1 jkn (7)

In fact

ek=yk-r=P(jk)ek-1+r-r=P(jk)ek-1

X1,X2,are sequences of independent random variables with the same distribution as X. For the convenience of notation, we denote Ek-1=EXkX1,X2,,Xk-1. That is to say, the conditional expectation is the condition on the first (k-1) iteration of the algorithm, thus obtaining

Ek-1ek22=Ek-1P(Xk)ek-122=
Ek-1<P(Xk)ek-1,P(Xk)ek-1>=Ek-1<ek-1,P(Xk)P(Xk)ek-1>=<ek-1,Ek-1[P(Xk)]ek-1>ek-12Im-A¯A¯TA¯F2ek-12
1-σmin2A¯F2ek-122

Among them, we applied the properties of expectation, Cauchy-Schwarz inequality and Lemma 1. Since

A¯+2=1σmin,κF(A¯)=A¯FA¯+2

we can obtain

Eek221-1κF2(A¯)ke022

Note that e0=b¯-r=b.

Theorem 2   In Algorithm 2, if the termination criterion of the randomized orthogonal projection algorithm is set as A¯Tyk2A¯Fyk2ε, it holds that yk-r2/yk2εκF(A¯), i.e.,b¯-ykb.

Proof   Assuming that the termination criterion is satisfied when some k>0. Set yk=r+w,wR(A¯), then

A¯Tyk2=A¯Tr+w2=A¯Tw2σminyk-r2

Since

A¯Tyk2εA¯Fyk2

we obtain that

yk-r2/yk2εA¯Fσmin

In particular, use again

A¯+2=1σmin,κF(A¯)=A¯FA¯+2

hence

yk-r2/yk2εκF(A¯)

that is

b¯-ykb

The stop criteria for Step 9 are determined on the basis of the following analysis. From the second part of the stop rule, we know

yk-r2εA¯F2σminxk2

Now, x̃ is the minimum norm solution of Eq.(5), then

A¯xk-x̃2A¯xk-(b¯-yk)2+b¯-yk-A¯x̃2εA¯Fxk2+r-ykεA¯Fxk2+εA¯F2σminxk2

Here we use the triangle inequality, the first part of the stop rule and b=A¯x̃. In particular,xk,x̃R(A¯T), it results that

xk-x̃2xk2εκF(A¯)(1+κF(A¯)) (8)

From Eq.(8), we can see that the relative error of RREK algorithm is bounded.

Theorem 3   gives the expected convergence rate of RREK algorithm (Algorithm 2).

Theorem 3   For any matrix A¯, right‑hand side vector b¯ and initial value x0=0, the sequence xk generated by Algorithm 2 converges to the minimum norm solution x̃ of Eq.(5)

Exk-x̃221-1κF2(A¯)k/2(1+2κ2(A¯))x̃2

Proof   For easy notation, denote

α=1-1/κF2AEk:=E[|i0,j0,i1,j1,,ik,jk]

According to Theorem 1, for each l0, we have

Eyl-r22αlb22b22 (9)

Fix a parameter k*:=k/2, after the k*-th iteration of Algorithm 2, it follows from reference13 that

Ek*-1xk*-x̃22αxk*-1-x̃22+r-yk*-122A¯F2 (10)

Indeed, randomized Kaczmarz algorithm is carried out on (A¯,b¯-yk*-1). Take the total expectation on both sides, due to the linear nature of the expectation, it holds that

Exk*-x̃22αExk*-1-x̃22+Er-yk*-122A¯F2
αExk*-1-x̃22+b22A¯F2
αk*x0-x̃22+l=0k*-2αlb22A¯F2
x̃22+l=0αlb22A¯F2 (11)

The last inequality is obtained from α<1,x0=0, and Eq.(11) can be simplified to

Exk*-x̃22x̃22+b22σmin2 (12)

using the fact l=0αl=11-α=κF2(A¯).

In addition, for each l0, we have

Er-yl+k*22αl+k*b22αk*b22 (13)

Now, for any k̃>0, similar considerations as to Eq.(11) implies that

Exk̃+k*-x̃22αExk̃+k*-1-x̃22+Er-yk̃+k*-122A¯F2
αk̃Exk*-x̃22+l=0k̃-1α(k̃-1)-lEr-yl+k*22A¯F2
αk̃Exk*-x̃22+αk*b22A¯F2l=0k̃-1αl
αk̃x̃22+b22/σmin2+αk*b22/σmin2=
αk̃x̃22+αk̃+αk*b22/σmin2
αk̃x̃22+αk̃+αk*κ2A¯x̃22 (14)

The last inequality is derived from b2σmaxx̃2.

Then, consider two cases: if k is even, set k̃=k*; otherwise, set k̃=k*+1. In both cases,(αk̃+αk*)2αk*. Therefore, Eq.(14) can be simplified as

Exk̃+k*-x̃22αk*1+2κ2(A¯)x̃22 (15)

3 Numerical Examples

In this section, the three numerical experiments are used to examine the RREK algorithm and compare it with REK algorithm. Relative error(RE) is used to measure the accuracy of the approximate solutions obtained by the two algorithms

RE=xk-xexact2xk2

where xk is the approximate solution of Eq.(1) derived from the REK and RREK at index k and xexact the exact solution of Eq.(1). The regularization parameter ω is determined by the discrepancy principle.L is selected as the first derivative operator. Choosing ε10-2 can meet the solution requirement. The running environment in the paper is MATLAB (2017b), and the processor is 1.6 GHz Intel Core i5.

Example 1 In this example, AR314×314 is generated by the problem14, the exact solution is xexact=sin0.010.01π,b̃=Axexact+δrandn(314,1)A¯R628×314δ is the noise level. We take δ=0.1%, 0.5%, 1%, 5%, respectively. Calculated with the two algorithms, and relative errors are obtained, as shown in Table 1. Fig.1 compares the REK and RREK solutions with the exact solution for the noise level δ=1%. At the same time, 10 sample points are selected at a medium distance from the reconstruction results, and the errors of the two methods at the sample points are compared, as shown in Fig.2.

Table 1 Relative errors of two reconstruction methods
δ/%0.10.515
REK 4.55e-2 6.06e-2 6.55e-2 1.18e-1
RREK 2.62e-2 3.49e-2 4.37e-2 8.24e-2

Fig.1 Original image and images reconstructed by two methods

Fig.2 Comparison of relative errors between two methods at 10 sample points

It can be seen from Table 1 that under the same noise level, the relative errors of RREK algorithm are smaller than the REK algorithm. In Fig.1 and Fig.2, we note that the RREK method gives a better approximation of the exact solution, indicating that RREK method is superior to REK method.

Example 2 Considering the phillips problem14, the first kind of Fredholm integral Equation is

-66ks,txsds=yt    -6<t<6

The kernel function and the right function are respectively

ks,t=1+cosπt-s3    t-s<3                  0                    t-s3
yt=(6-t)1+12cosπt3+92πsinπt3

The exact solution is

xt=1+cosπt3    t<3              0            t3

The integral equation is discretized into a matrix A with order 1 000, then b˜=Axexact+δrandn(1 000,1), and L is usually a discrete approximation of some derivative operators. Fig.3 shows the solutions of REK and RREK at δ= 1% with relative errors of 0.077 5 and 0.030 8, respectively. As can be seen from Fig.4, the iterative error of RREK algorithm is smaller than that of REK algorithm.

Fig.3 Comparison of REK and RREK solutions with exact solution

Fig.4 Relationship between the relative errors and the iterations of two algorithms

Example 3 Consider the two-dimensional image restoration problem15. The most common fuzzy function is the Gaussian impulse function, which can be described by the following symmetric banded Toeplitz matrix

(Tσ)ij=e-12i-jσ2i-j<band0other

where σ controls the shape of the Gaussian pulse function, and the larger σ is, the more ill-posed the problem is. In this example, the real image X is 100×100, then the projection operator AR10 000×10 000 is a symmetric double block Toeplitz matrix, xexactR10 000. Meanwhile, add 1% Gaussian noise, set σ=1 and band=5. RREK algorithm and REK algorithm are used to reconstruct the image, and the restoration effects are as follows.

Fig.5(a) is the original image of “Lena”, and Fig.5(b) is the image polluted by blur and noise. The relative error of Fig.5(c) recovered by REK algorithm is 12.95%, and that of Fig.5(d) recovered by RREK algorithm is 10.94%.

Fig.5 Original, blurred and restored “Lena” images

Fig.6(a) is the original image of “house”, Fig.6(b) is the image polluted by blur and noise, Fig.6(c) is the image recovered by REK algorithm with a relative error of 7.38%, and Fig.6(d) is the image recovered by RREK algorithm with a relative error of 5.43%.

Fig.6 Original, blurred and restored “house” images

By observing the relative errors and image reconstruction effects of the two methods, the relative errors of RREK algorithm are always smaller than those of REK algorithm, and the recovered images are smoother. Therefore, RREK algorithm is efficient and superior to REK algorithm.

4 Conclusions

Randomized extended Kaczmarz algorithm based on Tikhonov regularization is proposed to solve the linear discrete ill-posed problem, and the convergence of the algorithm is analyzed. Numerical experiments show that the algorithm is superior to the randomized extended Kaczmarz algorithm. In the numerical experiments, the regularization matrix is the first derivative matrix. Better results may be obtained by appropriately adjusting the selection of L, which will be studied later.

Contributions Statement

Ms. LIU Fengming designed the study, provided the idea, wrote the manuscript. Prof. WANG Zhengsheng complied the models and conducted the analysis. Ms. YANG Siyu contributed to the discussion and background of the study. Prof. XU Guili contributed to the date analysis. All authors commented on the manuscript draft and approved the submission.

Acknowledgements

This work was supported by the National Natural Science Foundations of China (Nos.11571171, 62073161, and 61473148).

Conflict of Interest

The authors declare no competing interests.

References

1

ENGL H WHANKE MNEUBAUER A. Regularization of inverse problems[M]. DordrechtKluwer Academic Publishers2000. [百度学术

2

GROETSCH C W. The theory of Tikhonov regularization for Fredholm equations of the first kind[M]. BostonPitman1984. [百度学术

3

KACZMARZ S. Approximate solution of systems of linear equations‍[J]. International Journal of Control1937576): 1-5. [百度学术

4

NATTERER F. The mathematics of computerized tomography‍[M]. PhiladelphiaUSA: Society for Industrial and Applied Mathematics2001. [百度学术

5

JIANG MWANG G. Convergence studies on iterative algorithms for image reconstruction‍[J]. IEEE Transactions on Medical Imaging2003225): 569-579. [百度学术

6

FRERIS N MZOUZIAS A. Fast distributed smoothing for network clock synchronization‍[C]//IEEE Conference on Decision and Control (CDC). [S.l.]: IEEE2012. [百度学术

7

STROHMER TVERSHYIN R. A randomized Kaczmarz algorithm with exponential convergence‍[J]. Journal of Fourier Analysis and Applications2009151): 262-278. [百度学术

8

NEEDELL D. Randomized Kaczmarz solver for noisy linear systems‍[J]. Bit Numerical Mathematics2010502): 395-403. [百度学术

9

PETRA SPOPA C. Single projection Kaczmarz extended algorithm‍[J]. Numerical Algorithms2015733): 1-16. [百度学术

10

ZOUZIAS AFRERIS N M. Randomized extended Kaczmarz for solving least squares‍[J]. SIAM Journal on Matrix Analysis and Applications2013342): 773-793. [百度学术

11

BAI Z ZWU W T. On convergence rate of the randomized Kaczmarz method‍[J]. Linear Algebra and Its Applications2018553252-269. [百度学术

12

BAI Z ZWU W T. On partially randomized extended Kaczmarz method for solving large sparse overdetermined inconsistentl inear systems‍[J]. Linear Algebra and Its Applications2019578225-250. [百度学术

13

SCHPFER FLORENZ D A. Linear convergence of the randomized sparse Kaczmarz method‍[J]. Mathematical Programming20191731): 509-536. [百度学术

14

HANSEN P C. Regularization tools version 4.0 for Matlab 7.3‍[J]. Numerical Algorithms2007462): 189-194. [百度学术

15

POPA CZDUNEK R. Kaczmarz extended algorithm for tomographic image reconstruction from limited data‍ [J]. Mathematics Computers Simulation2004656): 579-598. [百度学术

WeChat

Mobile website