Friday, September 20, 2019
Direct and iterative method
Direct and iterative method INTRODUCTION TO DIRECT AND ITERATIVE METHOD Many important practical problems give rise to systems of linear equations written as the matrix equation Ax = c, where A is a given n Ãâ" nnonsingular matrix and c is an n-dimensional vector; the problem is to find an n-dimensional vector x satisfying equation . Such systems of linear equations arise mainly from discrete approximations of partial differential equations. To solve them, two types of methods are normally used: direct methods and iterative methods. Directapproximate the solution after a finite number of floating point operations. Since computer floating point operations can only be obtained to a given precision, the computed solution is usually different from the exact solution. When a square matrix A is large and sparse, solving Ax = c by direct methods can be impractical, and iterative methods become a viable alternative. Iterative methods, based on splitting A into A = MâËâN, compute successive approximations x(t) to obtain more accurate solutions to a linear system at each iteration step t. This process can be written in the form of the matrix equation x(t) = Gx(tâËâ1) + g, where an n Ãâ" n matrix G = MâËâ1N is the iteration matrix. The iteration process is stopped when some predefined criterion is satisfied; the obtained vector x(t) is an approximation to the solution. Iterative methods of this form are called linear stationary iterative methods of the first degree. The method is of the first degree because x(t) depends explicitly only on x(tâËâ1) and not on x(tâËâ2), . . . , x(0). The method is linear because neither G nor g depends on x(tâËâ1), and it is stationary because neither G nor g depends on t. In this book, we also consider linear stationary iterative methods of the second degree, represented by the matrix equation x(t) = Mx(tâËâ1) âËâ Nx(tâËâ2) + h. HISTORY OF DIRECT AND ITERATIVE METHOD ÃË Direct methods to solve linear systems Direct methods for solving the linear systems with the Gauss elimination method is given byCarl Friedrich Gauss (1777-1855). Thereafter the Choleski gives method for symmetric positive definite matrices. ÃË Iterative methods for non-linear equations The Newton_Raphson method is an iterative method to solve nonlinear equations. The method is defined byIsaac Newton (1643-1727)andJoseph Raphson (1648-1715). ÃË Iterative methods for linear equations The standard iterative methods, which are used are the Gauss-Jacobi and the Gauss-Seidel method.Carl Friedrich Gauss (1777-1855)is a very famous mathematician working on abstract and applied mathematics.Carl Gustav Jacob Jacobi (1804-1851)is well known for instance for the Jacobian the determinant of the matrix of partial derivatives. He has also done work on iterative methods leading to the Gauss-Jacobi method. Another iterative method is the Chebyshev method. This method is based on orthogonal polynomials bearing the name ofPafnuty Lvovich Chebyshev (1821-1894). The Gauss-Jacobi and Gauss-Seidel method use a very simple polynomial to approximate the solution. In the Chebyshev method an optimal polynomial is used. DIRECT AND ITERATIVE METHOD Direct methods compute the solution to a problem in a finite number of steps. These methods would give the precise answer if they were performed ininfinite precision arithmetic. Examples includeGaussian elimination, theQRfactorization method for solvingsystems of linear equations, and thesimplex methodoflinear programming. In contrast to direct methods,iterative methodsare not expected to terminate in a number of steps. Starting from an initial guess, iterative methods form successive approximations thatconvergeto the exact solution only in the limit. Aconvergence criterionis specified in order to decide when a sufficiently accurate solution has (hopefully) been found. Even using infinite precision arithmetic these methods would not reach the solution within a finite number of steps (in general). Examples includeNewtons method, thebisection method, andJacobi iteration. In computational matrix algebra, iterative methods are generally needed for large problems. Iterative methods are more common than direct methods in numerical analysis. Some methods are direct in principle but are usually used as though they were not, e.g.GMRESand theconjugate gradient method. For these methods the number of steps needed to obtain the exact solution is so large that an approximation is accepted in the same manner as for an iterative method. In the case of asystem of linear equations, the two main classes of iterative methods are thestationary iterative methods, and the more generalKrylov subspacemethods. ÃË Stationary iterative methods Stationary iterative methods solve a linear system with anoperatorapproximating the original one; and based on a measurement of the error (the residual), form acorrection equationfor which this process is repeated. While these methods are simple to derive, implement, and analyse, convergence is only guaranteed for a limited class of matrices. Examples of stationary iterative methods are theJacobi method,Gaussââ¬âSeidel methodand theSuccessive over-relaxation method. ÃË Krylov subspace methods Krylov subspacemethods form anorthogonal basisof the sequence of successive matrix powers times the initial residual (theKrylov sequence). The approximations to the solution are then formed by minimizing the residual over the subspace formed. The prototypical method is theconjugate gradient method(CG). Other methods are the generalized minimal residual method and the biconjugate gradient method EXAMPLE OF DIRECT METHOD GAUSS ELIMINATION METHOD :- Inlinear algebra,Gaussian eliminationmethod is analgorithmfor solvingsystems of linear equations, finding therankof amatrix, and calculating the inverse of aninvertible square matrix. Gaussian elimination is named after German mathematician and scientistCarl Friedrich Gauss. Elementary row operationsare used to reduce a matrix torow echelon form.Gaussââ¬âJordan elimination, an extension of this algorithm, reduces the matrix further toreduced row echelon form. Gaussian elimination alone is sufficient for many applications. EXAMPLE Suppose that our goal is to find and describe the solution(s), if any, of the followingsystem of linear equations: The algorithm is as follows :eliminate x from all equations below L1 and then eliminate y from all equations below L2 .This will form a triangular form.Using the back substitution Each unknown can be solved . In the example, x is eliminated from l2 by adding 3/2L1to L2. X is then eliminatedmfrom L3 by adding L1 to L3 The result is: Nowyis eliminated fromL3by addingâËâ 4L2toL3: The result is: This result is a system of linear equations in triangular form, and so the first part of the algorithm is complete.The second part, back-substitution, consists of solving for the unknowns in reverse order. It can be seen that Then,zcan be substituted intoL2, which can then be solved to obtain Next,zandycan be substituted intoL1, which can be solved to obtain The system is solved. Some systems cannot be reduced to triangular form, yet still have at least one valid solution: for example, ifyhad not occurred inL2andL3after the first step above, the algorithm would be unable to reduce the system to triangular form. However, it would still have reduced the system toechelon form. In this case, the system does not have a unique solution, as it contains at least onefree variable. The solution set can then be expressed parametrically . In practice, one does not usually deal with the systems in terms of equations but instead makes use of theaugmented matrix(which is also suitable for computer manipulations). The Gaussian Elimination algorithm applied to theaugmented matrixof the system above, beginning with:which, at the end of the first part of the algorithm That is, it is inrow echelon form. At the end of the algorithm, if theGaussââ¬âJordan eliminationis applied: That is, it is inreduced row echelon form, or row canonical form. EXAMPLE OF ITERATIVE METHOD OF SOLUTION A. JACOB METHOD :- The Jacobi method is a method of solving amatrix equationon a matrix that has no zeros along its main diagonal (Bronshtein and Semendyayev 1997, p.892). Each diagonal element is solved for, and an approximate value taken in. The process is then iterated until it converges. This algorithm is a stripped-down version of theJacobi transformationmethod ofmatrix diagonalization. The Jacobi method is easily derived by examining each of the equations in thelinear system of equationsAx=b in isolation. If, in theith equation solve for the value ofwhile assuming the other entries ofremain fixed. This gives which is the Jacobi method. In this method, the order in which the equations are examined is irrelevant, since the Jacobi method treats them independently. The definition of the Jacobi method can be expressed withmatricesas B. Stationary Iterative Methods Iterative methods that can be expressed in the simple form Where neighter B nor c depend upon the iterative count k) are called stationary iterative method. The four main stationary iterative method : the Jacobi method, the Gauss Seidel method ,Successive Overrelaxation method and the symmetric Successive Overrelaxation method C. The Gauss-Seidel Method We are considering an iterative solution to the linear system where is ansparse matrix,xandbare vectors of lengthn, and we are solving forx. Iterative solvers are an alternative to direct methods that attempt to calculate an exact solution to the system of equations. Iterative methods attempt to find a solution to the system of linear equations by repeatedly solving the linear system using approximations to the vector. Iterations continue until the solution is within a predetermined acceptable bound on the error. Iterative methods for general matrices include the Gauss-Jacobi and Gauss-Seidel, while conjugate gradient methods exist for positive definite matrices. Use of iterative methods is the convergence of the technique. Gauss-Jacobi uses all values from the previous iteration, while Gauss-Seidel requires that the most recent values be used in calculations. The Gauss-Seidel method has better convergence than the Gauss-Jacobi method, although for dense matrices, the Gauss-Seidel method is sequential. The convergence of the iterative method must be examined for the application along with algorithm performance to ensure that a useful solution to can be found. The Gauss-Seidel method can be written as: where:à ¯ is theunknown in during theiteration,and, is the initial guess for theunknown in, is the coefficient ofin therow andcolumn, is thevalue in. or where:à ¯ K(k)is theiterative solution to is the initial guess atx Dis the diagonal ofA Lis the of strictly lower triangular portion ofA Uis the of strictly upper triangular portion ofA bis right-hand-side vector. EXAMPLE. 101âËâx2+ 23= 6, âËâx1+ 112âËâx3+ 34= 25, 21âËâx2+ 103âËâx4= âËâ 11, 32âËâx3+ 84= 15. Solving forx1,x2,x3andx4gives: x1=x2/ 10 âËâx3/ 5 + 3 / 5, x2=x1/ 11 +x3/ 11 âËâ 34/ 11 + 25 / 11, x3= âËâx1/ 5 +x2/ 10 +x4/ 10 âËâ 11 / 10, x4= âËâ 32/ 8 +x3/ 8 + 15 / 8 Suppose we choose(0,0,0,0)as the initial approximation, then the first approximate solution is given by x1= 3 / 5 = 0.6, x2= (3 / 5) / 11 + 25 / 11 = 3 / 55 + 25 / 11 = 2.3272, x3= âËâ (3 / 5) / 5 + (2.3272) / 10 âËâ 11 / 10 = âËâ 3 / 25 + 0.23272 âËâ 1.1 = âËâ 0.9873, x4= âËâ 3(2.3272) / 8 + ( âËâ 0.9873) / 8 + 15 / 8 = 0.8789. x1 x2 x3 x4 0.6 2.32727 âËâ 0.987273 0.878864 1.03018 2.03694 âËâ 1.01446 0.984341 1.00659 2.00356 âËâ 1.00253 0.998351 1.00086 2.0003 âËâ 1.00031 0.99985 The exact solution of the system is (1,2,-1,1) APPLICATION OF DIRECT AND ITERATIVE METHOD OF SOLUTION FRACTIONAL SPLITING METHOD OF FIRST ORDER FOR LINEAR EQUATION First we describe the simplest operator-splitting, which is calledsequential operator-splitting, for the following linear system of ordinary differential equations:(3.1)where the initial condition is. The operators and are linear and bounded operators in a Banach space The sequential operator-splitting method is introduced as a method that solves two subproblems sequentially, where the different subproblems are connected via the initial conditions. This means that we replace the original problem with the subproblemswhere the splitting time-step is defined as. The approximated solution is. The replacement of the original problem with the subproblems usually results in an error, calledsplitting error. The splitting error of the sequential operator-splitting method can be derived as whereis the commutator ofAandB The splitting error iswhen the operatorsA andB do not commute, otherwise the method is exact. Hence the sequential operator-splitting is called thefirst-order splitting method. THE ITERATIVE SPLITING The following algorithm is based on the iteration with fixed splitting discretization step-size. On the time intervalwe solve the following subproblems consecutively for:(4.1)where is the known split approximation at the time level. We can generalize the iterative splitting method to a multi-iterative splitting method by introducing new splitting operators, for example, spatial operators. Then we obtain multi-indices to control the splitting process; each iterative splitting method can be solved independently, while connecting with further steps to the multi-splitting method
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment
Note: Only a member of this blog may post a comment.