We study the smooth LU decomposition of a given analytic functional A-matrix A(A) and its block-analogue. Sufficient conditions for the existence of such matrix decompositions are given, some differentiability about certain elements arising from them are proved, and several explicit expressions for derivatives of the specified elements are provided. By using these smooth LU decompositions, we propose two numerical methods for computing multiple nonlinear eigenvalues of A(A), and establish their locally quadratic convergence properties. Several numerical examples are provided to show the feasibility and effectiveness of these new methods.
Signal and image restoration problems are often solved by minimizing a cost function consisting of an l2 data-fidelity term and a regularization term. We consider a class of convex and edge-preserving regularization functions. In specific, half-quadratic regularization as a fixed-point iteration method is usually employed to solve this problem. The main aim of this paper is to solve the above-described signal and image restoration problems with the half-quadratic regularization technique by making use of the Newton method. At each iteration of the Newton method, the Newton equation is a structured system of linear equations of a symmetric positive definite coefficient matrix, and may be efficiently solved by the preconditioned conjugate gradient method accelerated with the modified block SSOR preconditioner. Our experimental results show that the modified block-SSOR preconditioned conjugate gradient method is feasible and effective for further improving the numerical performance of the half-quadratic regularization approach.
The Hermitian and skew-Hermitian splitting (HSS) method is an unconditionally convergent iteration method for solving large sparse non-Hermitian positive definite system of linear equations. By making use of the HSS iteration as the inner solver for the Newton method, we establish a class of Newton-HSS methods for solving large sparse systems of nonlinear equations with positive definite Jacobian matrices at the solution points. For this class of inexact Newton methods, two types of local convergence theorems are proved under proper conditions, and numerical results are given to examine their feasibility and effectiveness. In addition, the advantages of the Newton-HSS methods over the Newton-USOR, the Newton-GMRES and the Newton-GCG methods are shown through solving systems of nonlinear equations arising from the finite difference discretization of a two-dimensional convection-diffusion equation perturbed by a nonlinear term. The numerical implemen- tations also show that as preconditioners for the Newton-GMRES and the Newton-GCG methods the HSS iteration outperforms the USOR iteration in both computing time and iteration step.
For large sparse non-Hermitian positive definite system of linear equations, we present several variants of the Hermitian and skew-Hermitian splitting (HSS) about the coefficient matrix and establish correspondingly several HSS-based iterative schemes. Theoretical analyses show that these methods are convergent unconditionally to the exact solution of the referred system of linear equations, and they may show advantages on problems that the HSS method is ineffective.
BAI Zhong-Zhi State Key Laboratory of Scientific/Engineering Computing,Institute of Computational Mathematics and Scientific/Engineering Computing,Academy of Mathematics and Systems Science,Chinese Academy of Sciences,P.O.Box 2719,Beijing 100080,China
We construct a modified Bernoulli iteration method for solving the quadratic matrix equation AX^2 + BX + C = 0, where A, B and C are square matrices. This method is motivated from the Gauss-Seidel iteration for solving linear systems and the ShermanMorrison-Woodbury formula for updating matrices. Under suitable conditions, we prove the local linear convergence of the new method. An algorithm is presented to find the solution of the quadratic matrix equation and some numerical results are given to show the feasibility and the effectiveness of the algorithm. In addition, we also describe and analyze the block version of the modified Bernoulli iteration method.
The restrictively preconditioned conjugate gradient (RPCG) method is further developed to solve large sparse system of linear equations of a block two-by-two structure. The basic idea of this new approach is that we apply the RPCG method to the normal-residual equation of the block two-by-two linear system and construct each required approximate matrix by making use of the incomplete orthogonal factorization of the involved matrix blocks. Numerical experiments show that the new method, called the restrictively preconditioned conjugate gradient on normal residual (RPCGNR), is more robust and effective than either the known RPCG method or the standard conjugate gradient on normal residual (CGNR) method when being used for solving the large sparse saddle point problems.